Get the latest tech news
Show HN: Real-time AI Voice Chat at ~500ms Latency
Have a natural, spoken conversation with AI! Contribute to KoljaB/RealtimeVoiceChat development by creating an account on GitHub.
This project lets you chat with a Large Language Model (LLM) using just your voice, receiving spoken responses in near real-time. Backend: Python 3.x, FastAPI Frontend: HTML, CSS, JavaScript (Vanilla JS, Web Audio API, AudioWorklets) Communication: WebSockets Containerization: Docker, Docker Compose Core AI/ML Libraries: RealtimeSTT(Speech-to-Text) RealtimeTTS(Text-to-Speech) transformers(Turn detection, Tokenization) torch/ torchaudio(ML Framework) ollama/ openai(LLM Clients) 🚀 Option A: Docker Installation (Recommended for Linux/GPU) This is the most straightforward method, bundling the application, dependencies, and even Ollama into manageable containers.
Or read this on Hacker News