Settings

Theme

A small tool I made for local LLMs: LLM-neofetch-plus

1 points by HFerrahoglu 2 days ago · 2 comments · 1 min read


Hey everyone!

I've had this in mind for a while, and I finally did it: it shows system information like regular NeoFetch, but I've added extra features for those using local LLMs (Ollama, llama.cpp, etc.).

For example: -How much VRAM does your GPU have, which model (NVIDIA, AMD, Intel, Apple M series)? -How many billions of parameters can your machine comfortably run (is 70B or 13B more sensible?) -Which GGUF quantization does what (Q4_K_M vs Q8_0 vs)? -Comparison of Ollama / llama.cpp / vLLM / LM Studio -Disk speed test + JSON/Markdown export

Simple installation: pip install llm-neofetch-plus

llm-neofetch -d 3 ← This is the detailed version, showing suggestions etc.

GitHub: https://github.com/HFerrahoglu/llm-neofetch-plus

If anyone tries it, could you tell me if you liked it or not, and what we should change? Thanks!

akssassin907 2 days ago

Running local LLMs on Apple Silicon has gotten surprisingly capable — the M-series chips handle models that used to require expensive GPU setups, so tools like this that actually speak to that hardware are welcome.

The quantization comparison is the feature I'd use most. It's one of those things that sounds simple but in practice nobody wants to dig through benchmarks just to figure out whether Q4 or Q8 is worth the extra memory on their specific machine.

Does it factor in what else is running in the background when estimating how much your machine can handle? That number can shift a lot depending on what else has memory tied up.

  • HFerrahogluOP a day ago

    No. The tool uses total hardware resources, not currently available ones.

    It displays current usage (via progress bars and "used" values), but the LLM model recommendations are based on raw hardware totals, not what's actually free after background processes.

    This is a reasonable design choice for a "system specs" tool—it's showing what the machine has, not what it has available right now. But you're right that it could misleadingly suggest a 70B model fits on a 24GB GPU when other apps are already using 4GB.

Keyboard Shortcuts

j
Next item
k
Previous item
o / Enter
Open selected item
?
Show this help
Esc
Close modal / clear selection