Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

It seems like a pointless discussion since DeepSeek uses Nvidia GPUs after all.


it uses a fractional amount of GPUs though.


As it says in the article, you are talking about a mere constant of proportionality, a single multiple. When you're dealing with an exponential growth curve, that stuff gets washed out so quickly that it doesn't end up matter all that much.

Keep in mind that the goal everyone is driving towards is AGI, not simply an incremental improvement over the latest model from Open AI.


Why do you assume that exponential growth curve is real?


Jevons Paradox states that increasing efficiency can cause an even larger increase in demand.


Their loss curve with the RL didn't level off much though, could be taken a lot further and scaled up to more parameters on the big nvidia mega clusters out there. And the architecture is heavily tuned to nvidia optimizations.


Which due to the Jevons Paradox may ultimately cause more shovels to be sold


"wait" I suspect we are all in a bit of denial.

When was the last time the US got their lunch ate in technology?

Sputnik might be a bit hyperbolic but after using the model all day and as someone who had been thinking of a pro subscription, it is hard to grasp the ramifications.

There is just no good reference point that I can think of.


Yep some CEO said they have 50K GPUs of the prior generation. They probably accumulated them through intermediaries that are basically helping nvidia sell to sanctioned parties by proxy


Deepseek was there side project. They had a lot of GPUs from their crypto mining project.

Then Ethereum turned off PoW mining, so they looked into other things to do with their GPUs, and started DeepSeek.


Mining crypto on H100s?




Consider applying for YC's Winter 2026 batch! Applications are open till Nov 10

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: