Settings

Theme

OpenLLaMA 7B Training Completed to 1T Tokens

huggingface.co

58 points by jncraton 3 years ago · 3 comments

Reader

mdaniel 3 years ago

be sure to read the warning in their repo: https://github.com/openlm-research/open_llama#loading-the-we...

> Please note that it is advised to avoid using the Hugging Face fast tokenizer for now, as we’ve observed that the auto-converted fast tokenizer sometimes gives incorrect tokenization

fancyfredbot 3 years ago

This is great. Based on the throughout of 2200 tokens/sec and the 1,000,000,000,000 tokens used to train this was at least $183k worth of compute (that's based on the three year committed use rate). And now we can have it for free!

Keyboard Shortcuts

j
Next item
k
Previous item
o / Enter
Open selected item
?
Show this help
Esc
Close modal / clear selection