Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Yes! I remember the "Obama stepping on the scale" example that was used in that article. Would love to know how GPT-4 performs on that test.



you mean this http://karpathy.github.io/2012/10/22/state-of-computer-visio...? Very funny to revisit. How primitive our tools were in comparison to now is astounding. It feels like the first flight of the Wright Brothers vs a jetliner. Imagenet was the new frontier. Simpler times...


I think the interesting thing here is the very, very surprising result that LLMs would be capable of abstracting the things in the second to last paragraph from the described experiences of amalgamated written human data.

It's the thing most people even in this thread don't seem to realize has emerged in research in the past year.

Give a Markov chain a lot of text about fishing and it will tell you about fish. Give GPT a lot of text about fishing and it turns out that it will probably learn how to fish.

World model representations are occuring in GPT. And people really need to start realizing there's already published research demonstrating that, as it goes a long way to explaining why the multimodal parts work.


Especially funny since the author, Andrej Karpathy, wrote at the end of the 2012 article that

>we are very, very far and this depresses me. What is the way forward? :( Maybe I should just do a startup

and was a founding member of OpenAI just a few years later in 2015


And he just rejoined them in February.


Didn't realize this was from 2012, but yes this is definitely what I was thinking of.


They say there are 3 mirrors in the scene but there are at least 5 - one which can only be seen indirectly through one of the other mirrors!




Consider applying for YC's Fall 2025 batch! Applications are open till Aug 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: