Settings

Theme

Ask HN: Compression Ratio of LLMs?

4 points by FileSorter a year ago · 3 comments · 1 min read

Reader

Does anyone know what the lossy compression ratio of a LLM is?

((Final .safetensors [GB]) / (Total Training Data [GB])) * 100 = ?

psyklic a year ago

I believe this wouldn't be meaningful, since any size LLM can be trained on any amount of data.

You could measure how well it memorizes via prediction accuracy on the training set, but this wouldn't indicate whether it generalizes well.

speedgoose a year ago

LLaMa 3.1 has been pre-trained on 15 trillion tokens, plus some more millions for the fine-tuning. About 60 terabytes.

https://github.com/meta-llama/llama-models/blob/main/models/...

The heaviest quantised LLaMa 3.1 8B is about 3.4GB.

So 0.005% compression rate, if you don't mind the intelligence of a heavily quantised 8B model.

schappim a year ago

OpenAI’s GPT-3 model (175B) has an archive size of about 350 GB, with training data estimated in the hundreds of terabytes, resulting in a highly compressed ratio.

Keyboard Shortcuts

j
Next item
k
Previous item
o / Enter
Open selected item
?
Show this help
Esc
Close modal / clear selection