Settings

Theme

Ask HN: On-Device vs. Cloud Based LLMs

2 points by zahirbmirza 2 months ago · 0 comments · 1 min read


According to Claude, "the underlying infrastructure uses shared compute resources — many users' requests are handled across a pool of GPUs"

What is the size of this pool, ie how many GPUs would it take for an individual user to be able to run their own equivalent today? Let's assume the LLM is fully downloadable.

I ask, because, if LLMs stop improving exponentially, surely soon enough we will ALL be able to run un-quantised local LLMs of sufficient quality for day to day tasks.

No comments yet.

Keyboard Shortcuts

j
Next item
k
Previous item
o / Enter
Open selected item
?
Show this help
Esc
Close modal / clear selection