Hacker News
new
|
past
|
comments
|
ask
|
show
|
jobs
|
submit
login
tuananh
on Feb 28, 2024
|
parent
|
context
|
favorite
| on:
The Era of 1-bit LLMs: ternary parameters for cost...
Major breakthrough in LLM scene. Achieve performance and perplexity equivalent to full FP16 models of same parameter size.
And you can fit 120B model with a single card 24GB VRAM. This is mind blowing.
cyanydeez
on Feb 28, 2024
[–]
I mean, it expands the hardware selection, but until there's models and leader boards etc, can't really say it's a break through.
fnordpiglet
on Feb 29, 2024
|
parent
[–]
I would assume a GPU isn’t specifically optimized for ternary computation and specialized accelerators would whip the pants off a GPU
Guidelines
|
FAQ
|
Lists
|
API
|
Security
|
Legal
|
Apply to YC
|
Contact
Search:
And you can fit 120B model with a single card 24GB VRAM. This is mind blowing.