Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

LLMs aren’t deterministic even with a seed.


Doesn't that depend on the implementation? There's a trade-off between performance and determinism for sure, but if determinism is what you want then it should be possible.


If you fix random seeds, disable dropout, and configure deterministic kernels, you can get reproducible outputs locally. But you still have to control for GPU non-determinism, parallelism, and even library version differences. Some frameworks (like PyTorch) have flags (torch.use_deterministic_algorithms(True)) to enforce this.


what if you set top_p=1, temperature=0, and always run it on the same local hardware


Horace He at Thinking Machines just dropped an awesome article describing exactly this: https://thinkingmachines.ai/blog/defeating-nondeterminism-in...

TL;DR: assuming you've squashed all regular non-determinism (itself a tall ask), you either need to ensure you always batch requests deterministically, or ensure all kernels are "batch invariant" (which is absolutely not common practice to do).


Maybe if you run it on CPU. (Maybe on GPU if all batching is disabled, but I wouldn't bet on it.)


cosmic wave will get you


Yes, that's the joke


This. I’m still amazed how many people don’t understand how this technology actually works. Even those you would think would have a vested interest in understanding it.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: