Hacker Newsnew | past | comments | ask | show | jobs | submitlogin
Ask HN: Pull the curtain back on Nvidia's CES keynote please
54 points by btbt 7 months ago | hide | past | favorite | 15 comments
I've spent 15 years building a career in robotics/software engineering along the "traditional path": writing software to solve problems, mastering fundamentals, and architecting complex systems. I stay on top of cutting-edge technologies to stay relevant, but I’ve mostly seen AI/LLMs as powerful tools for augmenting specific tasks rather than fundamentally reshaping engineering as a discipline.

However, NVIDIA’s recent CES keynote was another data point in a growing list challenging that view. While yes, primarily a sales pitch, the vision they presented suggests:

- AI can solve any problem across modalities—just feed it data.

- The future of engineering focuses on data gathering and AI integration, moving away from first-principles problem-solving.

While the presentation was compelling, my consumer experience with AI/LLMs has been vastly different:

- Great for suggestive tasks but unreliable for authoritative answers.

- Good for prototyping and scaffolding, but major rewrites are often needed for production-ready code.

- Image generation impresses but often includes noticeable flaws.

- Hallucinations across all modalities remain a major issue.

So, there’s this gap between the vision of AI/LLMs as a revolutionary paradigm and my practical experiences. I lack the resources to test these claims in depth, so I’m turning to HN for firsthand insights—not sales pitches or market hype.

If you’re using NVIDIA’s AI tools (like NeMo, Omniverse, COSMOS, etc.) or broadly working with LLMs/multi-modal AI in your daily work or business, I’d love to know:

- What are you actually achieving with them?

- What are their limitations? Where do they fall short?

- What's the reliability of the results? Demo-worthy Deployment-ready for high-stakes environments?

- Are the challenges you encounter just a matter of “more compute/money,” or are they fundamental barriers?

- Do these tools help you reach production faster compared to traditional methods?

- Anything else you think may be interesting.

I’m looking for honest, real-world insights from those at the front lines of these technologies.



Agree with all your points on the real world consumer experience.

* I would never assume the AI answer to a consequential problem to be authoritative, unless it shows me the source and I can click on the link to verify the source and the data presented (search engine use case).

* Rewrites with AI are bug-prone and often produce hard to trace bugs to the seemingly correct nature of these bugs. Generating the scaffolding works super well.

* Images are often too smooth, videos too robotic and rhythmic, water too shiny, etc. Trained eyes can easily distinguish between AI and real.

* Hallucinations are commonplace.


> it shows me the source and I can click on the link to verify the source ... (search engine use case)

To me, this is exactly a search engine. I type my query into 2005-2015 Google, I scan the page summaries under the links to see the answer, and click the best-looking result to confirm or read the details. Occasionally you need to re-word your query to get the answer you're looking for. Sometimes I don't bother clicking through because the answer is right there.

I don't really care that I can use plainer English with an "AI"; I'd be happier if I just got 2010 Google back. But sadly, it's gone.


> Images are often too smooth, videos too robotic and rhythmic, water too shiny, etc. Trained eyes can easily distinguish between AI and real.

That's likely to get better. Last year, consistently getting fingers and arms right was tough. This year, there are AI-generated violin playing videos.

> I would never assume the AI answer to a consequential problem to be authoritative, unless it shows me the source and I can click on the link to verify the source and the data presented (search engine use case).

That remains the elephant in the room - the tendency to make up fake answers. Until that's fixed, LLMs are only useful for problems where the cost of such errors is an externality, dumped on the consumer.


> > I would never assume the AI answer to a consequential problem to be authoritative, unless it shows me the source and I can click on the link to verify the source and the data presented (search engine use case).

> That remains the elephant in the room - the tendency to make up fake answers. Until that's fixed, LLMs are only useful for problems where the cost of such errors is an externality, dumped on the consumer.

That’s one of fears. The general public and politicians alike will trust AI without scrutiny. We’ve already seen examples of judges relying on flawed software, with devastating outcomes for innocent people. With the rapid push and widespread enthusiasm for AI, a darker future looms if these problems aren’t addressed.


I don't think the elephant can be solved by a tweak to LLMs. Producing a statistically-likely continuation of a pattern is what they do; there is no encoding of the world, just an encoding of language and image data.

A general crossing of that gap is, dare I say it, a problem requiring real intelligence.


> I don't think the elephant can be solved by a tweak to LLMs.

I doubt that too. But solving it is essential to the valuations of OpenAI and NVidia.


Wondering whether we will see some combination with Cyc at some point (which tried to solve the „encoding of the world“ problem)


After Doug's recent passing, I sincerely doubt it.


Related ongoing thread:

Jensen Huang keynote at CES 2025 [video] - https://news.ycombinator.com/item?id=42618595 - Jan 2025 (65 comments)


From the consumer perspective DLSS upscaling and frame prediction both had a lot of introduced artefacts that made them less than ideal even though they do improve performance quite a bit. This generation improves their accuracy and more heavily leans on the technologies to continue performance improvements. AMD is also making the same investments in their silicon putting considerable space to their AI cores and Ray tracing and not much at all to prior compute or rasterisation.

They are either right and its just a matter of more data and more compute thrown at the problem and its going to get indistinguishable from pixels more traditionally rendered or they are going to waste considerable silicon on the problem and it never becomes convincing. Even though its got problems DLSS3 has been quite popular for gamers and below perfect in these example are fine the errors aren't very consequential.

I don't know where this goes. I do know each generation of AI has improved quite a lot and we no longer talk about the turing test, we definitely took a jump but there remains a lot of hard engineering problems in every domain of AI to make it function as we want it to. Feels to me like a lot of these generators are in the uncanny valley, they are making the sort of errors that are weird and creepy but the thing is about that valley is it hides a lot of the progress being made.


I worked as an applied ML researcher for a while, so I'll give this a shot.

"- AI can solve any problem across modalities—just feed it data." - a large chunk of my time in ML is spent on data. I can't emphasize this enough - obtaining large amounts of quality data is a primary challenge with any sort of ML task. This might get easier with time, but will remain a challenge.

The corollary is that niche applications (and thus good fundamentals) are still important.

"- Are the challenges you encounter just a matter of “more compute/money,” or are they fundamental barriers?" - Well, there's a spectrum. Hallucinations are inherent to ML models - I don't think anybody has cracked ML model confidence estimation, and plenty have tried.

A slew of current limitations around LLMs stem from limited context windows. That is "only" inherent to the Transformer architecture (and there is some ongoing work on alternatives such as Mamba).

I think that "agents" and deep integration with computer interfaces will have some interesting automations come out of it.


I'd add to your list of problems: Publicly offered AIs are tuned to present a puritanical, sexless, inoffensive view of the world aligned with "the man" and kowtowing to corporate america's rules.


Nvidia robotic tools were somewhat convoluted last time I checked them couple years ago. I've seen people train robotic control systems with their sim, but for me it seems that the correct way is to set them as reference designs and reimplement in open source setting with constant community interest. Nvidia hardware and software reminds me military-grade stuff in their engineering approach, which may be sound for robotics. World models and simulators still in the state that doesn't require multibillion dollars to make progress in open setting.


It's interesting to hear you've had similar experience.

I used Isaac Sim for research work a few years ago (https://github.com/qcr/benchbot) and, although its capabilities were head and shoulders above the rest, its usability was amazingly convoluted. Even simple things became week-long dives into undocumented blind digging.

Part of why I'm so interested in the experiences of people actually using these tools for something productive.


I am making software for myself to learn, and to help my kids. I am using AI to essentially make A LOT of language exercises. It's really, really good at that. And learning is a lot more fun if you're creative with prompts.

I made javascript for a range of question types (things like fill-in, multiple-choice, ...) and have AI use that to e.g. generate short stories where you have to complete the verbs. Or replace some english words with german ones ... that sort of stuff.

Oh, and any time I need either tests or do something to a large range of variables, do what needs to be done to the first variable, copy over (using the reference) the list of fields, comment out the list of fields, and ask AI to suggest what to do. Usually only needs 1 or 2 changes.

And yes, I've noticed the hallucination. If you ask AI to correctly do a scatter-gather parallel processing in Go ... it's incredible how many errors it makes and it's infuriating how you have to explain every one of it's errors again and again. You have it output a basic structure, because that's still fast, and then rewrite the whole thing. I think it still gains me a bit of time ... but I see the point.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: