Most active commenters
  • brucethemoose2(5)
  • smiley1437(5)
  • LoganDark(4)

←back to thread

899 points georgehill | 15 comments | | HN request time: 1.468s | source | bottom
Show context
nivekney ◴[] No.36216106[source]
On a similar thread, how does it compare to Hippoml?

Context: https://news.ycombinator.com/item?id=36168666

replies(1): >>36216469 #
1. brucethemoose2 ◴[] No.36216469[source]
We don't necessarily know... Hippo is closed source for now.

Its comparable to Apache TVM's vulkan in speed on cuda, see https://github.com/mlc-ai/mlc-llm

But honestly, the biggest advantage of llama.cpp for me is being able to split a model so performantly. My puny 16GB laptop can just barely, but very practically, run LLaMA 30B at almost 3 tokens/s, and do it right now. That is crazy!

replies(1): >>36217701 #
2. smiley1437 ◴[] No.36217701[source]
>> run LLaMA 30B at almost 3 tokens/s

Please tell me your config! I have an i9-10900 with 32GB of ram that only gets .7 tokens/s on a 30B model

replies(3): >>36217877 #>>36217992 #>>36219745 #
3. LoganDark ◴[] No.36217877[source]
> Please tell me your config! I have an i9-10900 with 32GB of ram that only gets .7 tokens/s on a 30B model

Have you quantized it?

replies(1): >>36218570 #
4. oceanplexian ◴[] No.36217992[source]
With a single NVIDIA 3090 and the fastest inference branch of GPTQ-for-LLAMA https://github.com/qwopqwop200/GPTQ-for-LLaMa/tree/fastest-i..., I get a healthy 10-15 tokens per second on the 30B models. IMO GGML is great (And I totally use it) but it's still not as fast as running the models on GPU for now.
replies(2): >>36219157 #>>36219874 #
5. smiley1437 ◴[] No.36218570{3}[source]
The model I have is q4_0 I think that's 4 bit quantized

I'm running in Windows using koboldcpp, maybe it's faster in Linux?

replies(2): >>36219174 #>>36219792 #
6. LoganDark ◴[] No.36219157{3}[source]
> IMO GGML is great (And I totally use it) but it's still not as fast as running the models on GPU for now.

I think it was originally designed to be easily embeddable—and most importantly, native code (i.e. not Python)—rather than competitive with GPUs.

I think it's just starting to get into GPU support now, but carefully.

7. LoganDark ◴[] No.36219174{4}[source]
> The model I have is q4_0 I think that's 4 bit quantized

That's correct, yeah. Q4_0 should be the smallest and fastest quantized model.

> I'm running in Windows using koboldcpp, maybe it's faster in Linux?

Possibly. You could try using WSL to test—I think both WSL1 and WSL2 are faster than Windows (but WSL1 should be faster than WSL2).

replies(1): >>36220358 #
8. brucethemoose2 ◴[] No.36219745[source]
I'n on a Ryzen 4900HS laptop with a RTX 2060.

Like I said, very modest

replies(1): >>36220337 #
9. brucethemoose2 ◴[] No.36219792{4}[source]
I am running linux with cublast offload, and I am using the new 3 bit quant that was just pulled in a day or two ago.
replies(2): >>36220323 #>>36222560 #
10. brucethemoose2 ◴[] No.36219874{3}[source]
Have you tried the most recent cuda offload? A dev claims they are getting 26.2ms/token (38 tokens per second) on 13B with a 4080.
11. smiley1437 ◴[] No.36220323{5}[source]
Thanks! I'll have to try the 3bit to see if that helps
12. smiley1437 ◴[] No.36220337{3}[source]
Are you offloading layers to the RTX2060?
replies(1): >>36221349 #
13. smiley1437 ◴[] No.36220358{5}[source]
I didn't know what WSL was, but now I do, thanks for the tip!
14. brucethemoose2 ◴[] No.36221349{4}[source]
Some of them, yeah. 17 layers iirc.
15. LoganDark ◴[] No.36222560{5}[source]
cuBLAS or CLBlast? There is no such thing as cublast