←back to thread

544 points tosh | 1 comments | | HN request time: 0.211s | source
Show context
jauntywundrkind ◴[] No.43464180[source]
Wish I knew better how to estimate what sized video card one needs. HuggingFace link says this is bfloat16, so at least 64GB?

I guess the -7B might run on my 16GB AMD card?

replies(4): >>43464207 #>>43464240 #>>43464303 #>>43464853 #
1. wgd ◴[] No.43464207[source]
You can run 4-bit quantized version at a small (though nonzero) cost to output quality, so you would only need 16GB for that.

Also it's entirely possible to run a model that doesn't fit in available GPU memory, it will just be slower.