Hacker News new | past | comments | ask | show | jobs | submit login

The problem is it can’t remember what it hasn’t written but the end result still makes sense, so there has to be some goal after parsing the initial context that the tokens are emitted towards to. This means there’s nothing stopping it from producing an explanation, it might be in there from the very start.



Not one goal though, but a set of goals. Otherwise the random sampling of tokens wouldn't result in it outputting meaningfully different jokes.

I also think it's safe to assume that the set of goals isn't fully resolved, but it's harder to "prove" that.


There's no goal. The tokens are computed one at a time from all the previous tokens.


One is orthogonal to the other.




Join us for AI Startup School this June 16-17 in San Francisco!

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: