

Well, do we know what the blockers are for Tesla?
I feel like when I watch videos of FSD on cars, the representation of the world on the screen is rather good.
Now given this datapoint of me watching maybe 30minutes of video in total, is the issue in:
a) creating the distance to obstacles in the surroundings from cameras or in:
b) reading street signs, road markings, stop lights etc, or in:
c) doing the right thing, given a correct set of data about the surroundings?
Lidar / Radar / Sonar would only help for a).
Or is it combination of all of them, and the (relatively) cheap sensor would at least eliminate a), so one could focus on b and c?
But that’s exactly the point.
if the virtual map they’re building from cameras is complete, correct and stable (and presumably some other criteria that I didn’t think of from the top of my head), then the cameras would be sufficient.
The underlying neural decision network can still fuck things up from a correct virtual world map.
Now, how good is the virtual world map in real world conditions?