←back to thread

577 points simonw | 8 comments | | HN request time: 0s | source | bottom
1. pulkitsh1234 ◴[] No.44723561[source]
Is there any website to see the minimum/recommended hardware required for running local LLMs? Much like 'system requirements' mentioned for games.
replies(5): >>44723575 #>>44724036 #>>44724407 #>>44724966 #>>44725488 #
2. GaggiX ◴[] No.44723575[source]
https://apxml.com/tools/vram-calculator

This one is very good in my opinion.

replies(1): >>44723775 #
3. jxf ◴[] No.44723775[source]
Don't think it has the GLM series on there yet.
4. knowaveragejoe ◴[] No.44724036[source]
If you have a HuggingFace account, you can specify the hardware you have and it will show on any given model's page what you can run.
5. CharlesW ◴[] No.44724407[source]
> Is there any website to see the minimum/recommended hardware required for running local LLMs?

LM Studio (not exclusively, I'm sure) makes it a no-brainer to pick models that'll work on your hardware.

6. qingcharles ◴[] No.44724966[source]
This can be a useful resource too:

https://www.reddit.com/r/LocalLLaMA/

7. svachalek ◴[] No.44725488[source]
In addition to the tools other people responded with, a good rule of thumb is that most local models work best* at q4 quants, meaning the memory for the model is a little over half the number of parameters, e.g. a 14b model may be 8gb. Add some more for context and maybe you want 10gb VRAM for a 14gb model. That will at least put you in the right ballpark for what models to consider for your hardware.

(*best performance/size ratio, generally if the model easily fits at q4 you're better off going to a higher parameter count than going for a larger quant, and vice versa)

replies(1): >>44726001 #
8. nottorp ◴[] No.44726001[source]
> maybe you want 10gb VRAM for a 14gb model

... or if you have Apple hardware with their unified memory, whatever the assholes soldered in is your limit.