←back to thread

684 points prettyblocks | 1 comments | | HN request time: 0.001s | source

I mean anything in the 0.5B-3B range that's available on Ollama (for example). Have you built any cool tooling that uses these models as part of your work flow?
Show context
computers3333 ◴[] No.42790399[source]
https://gophersignal.com – I built GopherSignal!

It's a lightweight tool that summarizes Hacker News articles. For example, here’s what it outputs for this very post, "Ask HN: Is anyone doing anything cool with tiny language models?":

"A user inquires about the use of tiny language models for interesting applications, such as spam filtering and cookie notice detection. A developer shares their experience with using Ollama to respond to SMS spam with unique personas, like a millennial gymbro or a 19th-century British gentleman. Another user highlights the effectiveness of 3B and 7B language models for cookie notice detection, with decent performance achieved through prompt engineering."

I originally used LLaMA 3:Instruct for the backend, which performs much better, but recently started experimenting with the smaller LLaMA 3.2:1B model.

It’s been cool seeing other people’s ideas too. Curious—does anyone have suggestions for small models that are good for summaries?

Feel free to check it out or make changes: https://github.com/k-zehnder/gophersignal

replies(3): >>42791453 #>>42809880 #>>42819063 #
jkmcf ◴[] No.42819063[source]
RSS plz?
replies(1): >>42824873 #
1. computers3333 ◴[] No.42824873[source]
Hey, thanks for checking out GopherSignal! RSS is a great idea—I’ll be starting on it this weekend. Appreciate the suggestion!