A fully local, privacy-first voice chat setup running on a single machine (tested on WSL2 with an NVIDIA RTX 4070 SUPER). You talk to an LLM through a web UI using your microphone, it responds with a cloned custom voice — all processing stays on your hardware. No cloud APIs required.
There is a fairly high latency on non-Mac setup, but it is usable if you are patient. You can privately voice chat with your favorite video game character now!
Note that this README is not 100% tested, the real setup has been evolving organically. It may be a bit chaotic too. You may try asking if you are stuck on something. I recommend you just work through it together with an AI assistant. E.g.:
- Open terminal and start
wsl(you will need WSL set up on Windows; on Linux, this should all also work, and even easier) - Copy this file to your home directory.
- Set up
nenvwith node-22.