Most active commenters

    ←back to thread

    Docker Model Runner

    (www.docker.com)
    100 points kordlessagain | 11 comments | | HN request time: 0.001s | source | bottom
    1. rockwotj ◴[] No.43677187[source]
    Looks exactly like ollama but built into Docker desktop? Anyone know of any differences?
    replies(4): >>43677209 #>>43677230 #>>43677457 #>>43678593 #
    2. krick ◴[] No.43677209[source]
    They imply it should be somehow optimized for apple silicon, but, yeah, I don't understand what this is. If docker can use GPU, well, it should be able to use GPU in any container that makes use of it properly. If (say) ollama as an app doesn't use it properly, but they figured a way to do it better, it would make more sense to fix ollama. I have no idea why this should be a different app than, well, the very docker daemon itself.
    replies(1): >>43677560 #
    3. gclawes ◴[] No.43677230[source]
    Aren't some of the ollama guys ex-Docker guys?
    replies(1): >>43688544 #
    4. ammo1662 ◴[] No.43677457[source]
    They are using OCI artifacts to package models, so you can use your own registry to host these models internally. However, I just can't see any improvement comparing with a simple FTP server. I don't think the LLM models can adopt hierarchical structures like Docker, and thus cannot leverage the benefits of layered file systems, such as caching and reuse.
    replies(2): >>43677720 #>>43680735 #
    5. mappu ◴[] No.43677560[source]
    All that work (AGX acceleration...) is done in llama.cpp, not ollama. Ollama's raison d'être is a docker-style frontend to llama.cpp, so it makes sense that Docker would encroach from that angle.
    6. remram ◴[] No.43677720[source]
    I think ollama uses OCI too? At least it's trying to. https://github.com/ollama/ollama/issues/914#issuecomment-195...
    replies(1): >>43681007 #
    7. blitzar ◴[] No.43678593[source]
    Hear me out here ... it's like docker, but with Ai <pause for gasps and applause>.

    Seems fair to raise 1bn at a valuation of 100bn. (Might roll the funds over into pitching Kubernetes, but with Ai next month)

    replies(1): >>43678869 #
    8. danparsonson ◴[] No.43678869[source]
    What they really need is a Studio Ghibli'd version of their logo
    9. jesserwilliams ◴[] No.43680735[source]
    It's not the only one using OCI to package models. There's a CNCF project called KitOps (https://kitops.org) that has been around for quite a bit longer. It solves some of the limitations that using Docker has, one of those being that you don't have to pull the entire project when you want to work on it. Instead, you can pull just the data set, tuning, model, etc.
    10. hobofan ◴[] No.43681007{3}[source]
    Yes, ollama also uses OCI, but currently only works with unauthenticated registries.
    11. rockwotj ◴[] No.43688544[source]
    yes