Get the latest tech news

Show HN: Real-time AI Voice Chat at ~500ms Latency


Have a natural, spoken conversation with AI! Contribute to KoljaB/RealtimeVoiceChat development by creating an account on GitHub.

This project lets you chat with a Large Language Model (LLM) using just your voice, receiving spoken responses in near real-time. Backend: Python 3.x, FastAPI Frontend: HTML, CSS, JavaScript (Vanilla JS, Web Audio API, AudioWorklets) Communication: WebSockets Containerization: Docker, Docker Compose Core AI/ML Libraries: RealtimeSTT(Speech-to-Text) RealtimeTTS(Text-to-Speech) transformers(Turn detection, Tokenization) torch/ torchaudio(ML Framework) ollama/ openai(LLM Clients) 🚀 Option A: Docker Installation (Recommended for Linux/GPU) This is the most straightforward method, bundling the application, dependencies, and even Ollama into manageable containers.

Get the Android app

Or read this on Hacker News

Read more on:

Photo of voice chat

voice chat

Photo of time AI

time AI

Photo of ~500ms latency

~500ms latency

Related news:

News photo

Batch data processing is too slow for real-time AI: How open-source Apache Airflow 3.0 solves the challenge with event-driven data orchestration

News photo

The Nintendo Switch 2 offers built-in voice chat and screensharing

News photo

VLC player demos real-time AI subtitling for videos