Settings

Theme

LLM in a Flash: Efficient Large Language Model Inference with Limited Memory

arxiv.org

12 points by keep_reading 2 years ago · 1 comment

Reader

dang 2 years ago

LLM in a Flash: Efficient LLM Inference with Limited Memory - https://news.ycombinator.com/item?id=38704982 - Dec 2023 (52 comments)

Keyboard Shortcuts

j
Next item
k
Previous item
o / Enter
Open selected item
?
Show this help
Esc
Close modal / clear selection