←back to thread

156 points Sean-Der | 1 comments | | HN request time: 0.223s | source

Alt link: https://mrchristmas.com/products/santas-magical-telephone

Video demo: https://www.youtube.com/watch?v=0z7QJxZWFQg

The first time I talked with AI santa and it responded with a joke I was HOOKED. The fun/nonsense doesn't click until you try it yourself. What's even more exciting is you can build it yourself:

libpeer: https://github.com/sepfy/libpeer

pion: https://github.com/pion/webrtc

Then go do all your fun logic in your Pion server. Connect to any Voice AI provider, or roll your own via Open Source. Anything is possible.

If you have questions or hit any roadblocks I would love to help you. I have lots of hardware snippets on my GitHub: https://github.com/sean-der.

Show context
architectonic ◴[] No.45574021[source]
How much computing power would one need to get this working completely local running a half decent llm fine tuned to sound like santa with all tts, stt and the pipecat inbetween?
replies(4): >>45574195 #>>45575432 #>>45576228 #>>45583058 #
1. kwindla ◴[] No.45583058[source]
This repo is one possible starting point for tinkering with local agents on macOS. I've got versions of this for NVIDIA platforms but I tend to gravitate to using LLMs that are too big to fit on most NVIDIA consumer cards.

https://github.com/kwindla/macos-local-voice-agents