Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

You are vastly underestimating the price decline. To cherrypick one article; in the first two years since GPT 3.5, inference price for the same amount of intelligence has decreased 10x per year according to a study by Andreessen Horowitz https://a16z.com/llmflation-llm-inference-cost/. So in a stark slowdown scenario, we could still see a 1000x decrease in the next 5 years.

Price deflation is not tied to Moore's right now because much of the performance gains are from model optimization, high bandwidth memory supply chains, and electrical capacity build out, not FLOP density.





True! I just know that model optimization gains are much less guaranteed than say, FLOP density, even though model optimization has so far provided way more gains than hardware advancements.

Part of me is optimistic that when the AI bubble bursts the excess data center capacity is going to be another force driving the cost of inference down.


> I just know that model optimization gains are much less guaranteed than say, FLOP density, even though model optimization has so far provided way more gains than hardware advancements.

Performance gained from model improvements has outpaced performance gained from hardware improvements for decades.


Haha, I love how delusional everyone is about AI.

Yeppers, when that bubble burst - that's hilarious. This is the kinda stuff grandkids won't believe someday.


> has decreased 10x per year according to a study by Andreessen Horowitz

I believe you but that's not exactly an unbiased source of information.




Consider applying for YC's Winter 2026 batch! Applications are open till Nov 10

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: