Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

> In total, the cerebellum contains 80% of all neurons!

Apples and oranges, but that's so reminiscent of MLPs in Transformers. A similarly large fraction of the weights in transformers come from the MLPs in each layer.



Well, MLPs is the actual neural network (the approximator), whereas the attention (the rest) is more of a text to relevant embedding extractor




Consider applying for YC's Fall 2025 batch! Applications are open till Aug 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: