Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

> So, LLMs face a regression on their latest proposed improvement.

Arguably a second regression, the first being cost, because COT improves performance by scaling up the amount of compute used at inference time instead of training time. The promise of LLMs was that you do expensive training once and then run the model cheaply forever, but now we're talking about expensive training followed by expensive inference every time you run the model.



To be fair they also advanced in the cost aspect with other models

gpt4o and 4o mini have a tenth and a hundredth of inference cost of gpt4 respectively




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: