(no title)
akssassin907 | 5 days ago
The quantization comparison is the feature I'd use most. It's one of those things that sounds simple but in practice nobody wants to dig through benchmarks just to figure out whether Q4 or Q8 is worth the extra memory on their specific machine.
Does it factor in what else is running in the background when estimating how much your machine can handle? That number can shift a lot depending on what else has memory tied up.
HFerrahoglu|4 days ago
It displays current usage (via progress bars and "used" values), but the LLM model recommendations are based on raw hardware totals, not what's actually free after background processes.
This is a reasonable design choice for a "system specs" tool—it's showing what the machine has, not what it has available right now. But you're right that it could misleadingly suggest a 70B model fits on a 24GB GPU when other apps are already using 4GB.