Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I'd bet it's a lot easier to write high reliability self driving software when you have multiple redundant sensors that don't all go to shit when it's foggy. There's probably a good reason that after a decade of r&d, every major player in this space with the exception of the one run by a narcissistic moron is using radar, lidar, and cameras in their sensor suite.


How are humans able to drive safely in the fog without redundant sensors?


Humans can't drive safely in the fog. They can drive safely on a road travelled often even with low visibility by using visual hints they need to "scan" at lower speed. No car taking a cam approach for self-driving is doing that btw.


And yet, conceptually, there's still no capacity humans have that camera-only autonomous vehicles can't be given, even if they don't have it right now, right?


Theory says so, facts paint a different picture where CV systems are foiled more often (in good visibility conditions) than the same situation where a human could be foiled with camouflage or trompe-l'oeil.

Something is missing, time will tell if training will produce more solid models.




Consider applying for YC's Fall 2025 batch! Applications are open till Aug 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: