Settings

Theme

Show HN: Selfhostllm.org – Plan GPU capacity for self-hosting LLMs

selfhostllm.org

7 points by erans 4 months ago · 3 comments · 1 min read

Reader

A simple calculator that estimates how many concurrent requests your GPU can handle for a given LLM, with shareable results.

eransOP 4 months ago

I also added a Mac version: https://selfhostllm.org/mac/ so you can know which models you can run on your Mac and get an estimated tokens/sec.

harshnigam 4 months ago

I see it doesn't take GPU performance into consideration when showing the estimates. H100 and A100 are performing the same. Am I doing it wrong?

atmanactive 4 months ago

Very useful, thanks. I'm missing a reset button though.

Keyboard Shortcuts

j
Next item
k
Previous item
o / Enter
Open selected item
?
Show this help
Esc
Close modal / clear selection