Most active commenters

    ←back to thread

    544 points tosh | 17 comments | | HN request time: 0.862s | source | bottom
    Show context
    simonw ◴[] No.43464227[source]
    Big day for open source Chinese model releases - DeepSeek-v3-0324 came out today too, an updated version of DeepSeek v3 now under an MIT license (previously it was a custom DeepSeek license). https://simonwillison.net/2025/Mar/24/deepseek/
    replies(5): >>43464375 #>>43464498 #>>43464686 #>>43465383 #>>43467111 #
    chaosprint ◴[] No.43464375[source]
    it seems that this free version "may use your prompts and completions to train new models"

    https://openrouter.ai/deepseek/deepseek-chat-v3-0324:free

    do you think this needs attention?

    replies(7): >>43464399 #>>43464480 #>>43464512 #>>43464616 #>>43464961 #>>43468548 #>>43470210 #
    1. huijzer ◴[] No.43464512[source]
    Since we are on HN here, I can highly recommend open-webui with some OpenAI-compatible provider. I'm running with Deep Infra for more than a year now and am very happy. New models are usually available within one or two days after release. Also have some friends who use the service almost daily.
    replies(7): >>43464718 #>>43465081 #>>43466430 #>>43466464 #>>43466949 #>>43469369 #>>43473139 #
    2. unquietwiki ◴[] No.43464718[source]
    I'm using open-webui at home with a couple of different models. gemma2-9b fits in VRAM on a NV 3060 card + performs nicely.
    replies(2): >>43465594 #>>43469350 #
    3. l72 ◴[] No.43465081[source]
    I too run openweb-ui locally and use deepinfra.com as my backend. It has been working very well, and I am quite happy with deepinfra's pricing and privacy policy.

    I have set up the same thing at work for my colleagues, and they find it better than openai for their tasks.

    replies(1): >>43468529 #
    4. zakki ◴[] No.43465594[source]
    What is the memory of your NV3060? 8GB?
    replies(1): >>43466243 #
    5. ngvjmfgb ◴[] No.43466243{3}[source]
    12GB (edit: that is what mine is)
    6. ◴[] No.43466430[source]
    7. totetsu ◴[] No.43466464[source]
    And it’s quite easy to set up a Cloudflare tunnel to make your open-webui instance accessible online too just you
    replies(1): >>43466484 #
    8. simonw ◴[] No.43466484[source]
    ... or a TailScale network. I've been leaving open-webui running on my laptop on my desk and then going out into the word and accessing it from my phone via TailScale, works great.
    replies(2): >>43466956 #>>43467521 #
    9. wkat4242 ◴[] No.43466949[source]
    Yeah OpenWebUI is great with local models too. I love it. You can even do a combo, send the same prompt to local and cloud and even various providers and compare the results.
    10. wkat4242 ◴[] No.43466956{3}[source]
    Yeah this sounds like the more secure option, you don't want to be dependent on a single flaw in a web service
    11. totetsu ◴[] No.43467521{3}[source]
    I would use tail scale. But I specifically want to use open web-ui from a place I can’t install a Tailscale client
    replies(1): >>43468733 #
    12. jychang ◴[] No.43468529[source]
    Yeah, openweb-ui is the best frontend for API queries. Everything seems to work well.

    I've tried LibreChat before, but the app is terrible at generating titles for chats instead of leaving it as "New Chat". Also it lacks a working Code Interpreter.

    13. fragmede ◴[] No.43468733{4}[source]
    where's that?
    14. mdp2021 ◴[] No.43469350[source]
    > performs nicely

    Do you have rough indication of token/s ?

    15. eurekin ◴[] No.43469369[source]
    I've tried using it, but it's browser tab seems to peg one core to 100% after some time. Anyone else experienced it?
    16. indigodaddy ◴[] No.43473139[source]
    Can open-webui update code on your local computer ala cursor etc?
    replies(1): >>43473514 #
    17. cess11 ◴[] No.43473514[source]
    It has a module system so maybe it can but it seems more people are using Aider or Continue for that. There's a bit of stitching things together regardless of whether you show your project to some SaaS or run local models but if you can manage a Linux system it'll be easy.

    Personally I heavily dislike the experience though, so I might not be the best one to answer.