Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

What did you expect?


I never know what to expect anymore. We live in a world where computers can describe paintings and write sonnets about them but a half-trillion dollar car company can't figure out how to parallel park with eight cameras.


DriveGPT as it hits 2 parked cars and runs over a dog: "Apologies for the confusion earlier. You are correct that the cars were in the planned local planner path..."


Sorry, yes, that was a dog.

Let me try again and this time I will definitely not hit anything.

Sorry, that was another dog.


"Was that another dog?"

BingDrive: I'm sorry, but I prefer not to continue this conversation.


It's like something right out of a Douglas Adams book.


That's a well known result from the last AI hype era.

https://en.wikipedia.org/wiki/Moravec%27s_paradox

Although, I'm not surprised AI can describe paintings, but I'm still pretty surprised it can generate them. How'd we get that to work!?


Yeah, and I'm still completely lost as to why resolution is such a limiting factor. If you know you're drawing a soccer ball why is a 512x512 soccer ball so much easier than a 1024x1024 soccer ball?


There are a few cases where people have used ChatGPT to generate SVG[0], with mostly unimpressive results. I'm sure sooner or later models will be developed specifically for creating vector drawings instead of raster, including with the ability to apply textures and gradients.

Also, the resolution of Stable Diffusion's output isn't much of a limitation if you're willing to use other tools to massage the output into something professional-quality. See [1]

[0] For example: https://praeclarum.org/2023/04/03/chatsvg.html

[1] https://news.ycombinator.com/item?id=35592847


That's because the decided they do not need lidar.


It's not lidar they need. BMW, Mercedes, Porsche, etc. All can park themselves almost perfectly every time. Teslas can't, and will take 5x as long to do it when they can, if the computer can even bother to recognise the space.

It's software. Original Teslas with AP1 better than Teslas own in house software on their latest AP.


Near term it’s a frustrating decision, but if these gpt4 vision LLMs are anything to go by it will prove to be the right decision in the long term.


Remember that "cameras" aren't as good as human perception because human eyes interact with the environment instead of being passive sensors. (That is, if you can't see something you can move your head.)

Plus we have ears, are under a roof so can't get rained on, are self cleaning, temperature regulating, have much better dynamic range, wear driving glasses…


And we still get into millions of accidents every year…


Which sounds like a lot until you realize 1) we drive over three trillion miles a year in the US, and 2) the majority of those accidents are concentrated to a fraction of all drivers. The median human driver is quite good, and the state of the art AI isn't even in the same galaxy yet.


I keep hearing this argument over and over, but I find it uncompelling. As a relatively young person with good vision, who has never been in an accident after many years of driving, and who doesn't make the kind of simple mistakes I've seen the absurd mistakes self-driving cars make and I would not trust my life to a self-driving car.

Asking people to accept a driverless car based on over-arching statistics is papering over some very glaring issues. For example, are most accidents in cars being caused by "average" drivers or are they young / old / intoxicated / distracted / bad vision? Are the statistics randomly distributed (e.g. any driver is just as likely as the next to get in accidents)? Because the driverless cars seem to have accidents at random in unpredictable ways, but human drivers can be excellent (no accidents, no tickets ever), or terrible (drive fast, tickets, high insurance, accidents, etc). The distribution of accidents among humans is not close to uniform, and is usually explainable. I wouldn't trust a poor human driver on a regular basis, nor would I trust an AI because I'm actually a much better driver than both (no tickets, no accidents, can handle complex situations the AI can't). Are the comparisons of human accidents being treated as homogenous (e.g. the chance of ramming full speed into a parked car the same as a fender-bender?). I see 5.8M car crashes anually, but deaths remain fairly low (~40k, .68%), vs 400 driverless accidents with ~20 deaths (5%), I'm not sure we're talking about the same type of accidents.

tl;dr papering over the complexity of driving and how good a portion of drivers might be by mixing non-homogenous groups of drivers and taking global statistics of all accidents and drivers to justify unreliable and relatively dangerous technology would be a strict downgrade for most good drivers (who are most of the population).


Why wouldn’t LIDAR in addition to computer vision with cameras be a strictly better idea?


It's all trade offs. I'm just spitballing here, but if you have limited resources, you can either spend cash/time on lidar or invest in higher-quality mass-produced optics, or better computer vision software. If you get to a functional camera-only system sooner, might everyone be better off as you can deploy it more rapidly.

Manufacturing capacity of lidar components might be limited.

Another might be reliability/failure modes. If the system relies on lidar, that's another component that can break (or brownout and produce unreliable inputs).

So in a vaccum, yea a lidar+camera system is probably better, but who knows with real life trade offs.

(again, I just made these up, I do not work on this stuff, but these are a few scenarios I can imagine)


Couldn't afford*


Sure, sure, but would it have killed them to drop in a few five dollar "don't hit this object" ultrasonic proximity sensors?


While ultrasonic sensors would be fine for parking, they don't have very good range so they aren't much help in avoiding, for example, crashing into stationary fire trucks or concrete lane dividers at freeway speeds.


Just disable autopilot 0.00001 seconds before impact and it becomes the driver's fault.


Don’t worry. That’s valuation has always been ZIRP nonsense and will fall with Musk’s addiction to publicly demonstrating his incompetence.


A 14-line poem with a consistent rhyme scheme and meter. Perhaps my request should have been more specific.


From my experimentation, LLMs tend to kind of suck at rhyme and meter, and all but the simplest types of poetry, so even if you'd specified it probably wouldn't have been able to deliver.

This is definitely something they could be trained to be much better at, but I guess it's hasn't been a priority.


GPT4 is surprisingly good at it, considering BPE tokenization means it shouldn't be able to do rhyme at all.


Has anyone tried using phonetic tokens instead of text? I'm curious if that would help with things like rhyming.




Consider applying for YC's Fall 2025 batch! Applications are open till Aug 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: