←back to thread

343 points LorenDB | 6 comments | | HN request time: 0.001s | source | bottom
Show context
tommica ◴[] No.44002018[source]
Sidetangent: why is ollama frowned upon by some people? I've never really got any other explanation than "you should run llama.CPP yourself"
replies(9): >>44002029 #>>44002150 #>>44002166 #>>44002486 #>>44002513 #>>44002621 #>>44004218 #>>44005337 #>>44006200 #
nicman23 ◴[] No.44002029[source]
cpp was just faster and with more features that is all
replies(1): >>44002169 #
1. cwillu ◴[] No.44002169[source]
cpp is the thing doing all the heavy lifting, ollama is just a library wrapper.

It'd be like if handbrake tried to pretend that they implemented all the video processing work, when it's dependent on libffmpeg for all of that.

replies(1): >>44004229 #
2. diggan ◴[] No.44004229[source]
> ollama is just a library wrapper.

Was.

This submission is literally about them moving away from being just a wrapper around llama.cpp :)

replies(1): >>44005522 #
3. buyucu ◴[] No.44005522[source]
no they are not. the submission uses ggml, which is llama.cpp
replies(1): >>44006311 #
4. diggan ◴[] No.44006311{3}[source]
I think you misunderstand how these pieces fit together. llama.cpp is library that ships with a CLI+some other stuff, ggml is a library and Ollama has "runners" (like an "execution engine"). Previously, Ollama used llama.cpp (which uses ggml) as the only runner. Eventually, Ollama made their own runner (which also uses ggml) for new models (starting with gemma3 maybe?), still using llama.cpp for the rest (last time I checked at least).

ggml != llama.cpp, but llama.cpp and Ollama are both using ggml as a library.

replies(1): >>44007638 #
5. cwillu ◴[] No.44007638{4}[source]
“The llama.cpp project is the main playground for developing new features for the ggml library” --https://github.com/ggml-org/llama.cpp

“Some of the development is currently happening in the llama.cpp and whisper.cpp repos” --https://github.com/ggml-org/ggml

replies(1): >>44010025 #
6. diggan ◴[] No.44010025{5}[source]
Yeah, those both makes sense. ggml was split from llama.cpp once they realized it could be useful elsewhere, so while llama.cpp is the "main playground", it's still used by others (including llama.cpp). Doesn't mean suddenly that llama.cpp is the same as ggml, not sure why you'd believe that.