←back to thread

684 points prettyblocks | 5 comments | | HN request time: 0.015s | source

I mean anything in the 0.5B-3B range that's available on Ollama (for example). Have you built any cool tooling that uses these models as part of your work flow?
Show context
mettamage ◴[] No.42784724[source]
I simply use it to de-anonymize code that I typed in via Claude

Maybe should write a plugin for it (open source):

1. Put in all your work related questions in the plugin, an LLM will make it as an abstract question for you to preview and send it

2. And then get the answer with all the data back

E.g. df[“cookie_company_name”] becomes df[“a”] and back

replies(4): >>42784789 #>>42785696 #>>42785808 #>>42788777 #
politelemon ◴[] No.42784789[source]
Could you recommend a tiny language model I could try out locally?
replies(1): >>42784953 #
1. mettamage ◴[] No.42784953[source]
Llama 3.2 has about 3.2b parameters. I have to admit, I use bigger ones like phi-4 (14.7b) and Llama 3.3 (70.6b) but I think Llama 3.2 could do de-anonimization and anonimization of code
replies(2): >>42785057 #>>42785333 #
2. OxfordOutlander ◴[] No.42785057[source]
+1 this idea. I do the same. Just do it locally using ollama, also using 3.2 3b
3. RicoElectrico ◴[] No.42785333[source]
Llama 3.2 punches way above its weight. For general "language manipulation" tasks it's good enough - and it can be used on a CPU with acceptable speed.
replies(1): >>42785773 #
4. seunosewa ◴[] No.42785773[source]
How many tokens/s?
replies(1): >>42792310 #
5. iamnotagenius ◴[] No.42792310{3}[source]
10-15t/s on 12400 with ddr5