←back to thread

156 points Sean-Der | 1 comments | | HN request time: 0s | source

Alt link: https://mrchristmas.com/products/santas-magical-telephone

Video demo: https://www.youtube.com/watch?v=0z7QJxZWFQg

The first time I talked with AI santa and it responded with a joke I was HOOKED. The fun/nonsense doesn't click until you try it yourself. What's even more exciting is you can build it yourself:

libpeer: https://github.com/sepfy/libpeer

pion: https://github.com/pion/webrtc

Then go do all your fun logic in your Pion server. Connect to any Voice AI provider, or roll your own via Open Source. Anything is possible.

If you have questions or hit any roadblocks I would love to help you. I have lots of hardware snippets on my GitHub: https://github.com/sean-der.

Show context
architectonic ◴[] No.45574021[source]
How much computing power would one need to get this working completely local running a half decent llm fine tuned to sound like santa with all tts, stt and the pipecat inbetween?
replies(4): >>45574195 #>>45575432 #>>45576228 #>>45583058 #
teaearlgraycold ◴[] No.45574195[source]
I started looking into this with a Pi 5. It seemed like it was not quite performant enough. But I'm not an expert with these things and maybe someone else could make it work. We definitely have the technology to pull this off in this form factor. It would just be really expensive (maybe $500) and might also get a little hot.
replies(1): >>45575145 #
1. Sean-Der ◴[] No.45575145[source]
If I was building it to be 'local only' I would run the inference on a remote host in my house.

Having a microcontroller in the phone is nice because it is WAY less likely to break. I love being able to flash a simple firmware/change things would fighting it too much.

Oh! Also I do all the 'WebRTC/AI dev' in the browser. When I get it working how I like, then do I switch over to doing the microcontroller stuff.