ChatTTS: Making Open-Source TTS Sound Conversational
Open-source TTS models could read text aloud, but conversational naturalness—realistic rhythm, laughter, pauses—remained elusive. ChatTTS addresses prosody control directly, with growing pains that reflect its ambitious scope. We examine the technical approach, competitive landscape, and real-world quality issues reported by 38K+ community members.

Open-source text-to-speech models can read text aloud. Making them sound like actual human conversation—with natural pauses, laughter breaks, and the messy emotional rhythm of real dialogue—requires solving prosody, and that's where most previous attempts fell short.
Prosody is the melody of speech: pitch changes, timing, stress patterns, the difference between "I didn't say he stole the money" and "I didn't say he stole the money." It's what separates reading a script from having a conversation. For chatbots, voice assistants, and accessibility tools, perfect diction with wooden delivery still sounds wrong because the emotional texture is missing.
ChatTTS approaches this by giving developers control over conversational elements—not just speed or pitch, but laughter timing, pause duration, and interjection placement. Instead of treating dialogue as sentences to be read, the model handles it as conversation to be performed.
What controllable speech elements actually mean
The 2noise team built ChatTTS around the assumption that dialogue TTS needs different controls than audiobook narration. You can specify where laughter should occur, how long pauses should last between thoughts, and how much emotional coloring to apply to phrases. The model generates audio that includes these prosodic markers naturally, rather than splicing them in post-processing.
This matters when you're building a customer service bot that needs to sound empathetic or an educational assistant that should pause for comprehension. Generic TTS models optimize for clarity; ChatTTS optimizes for conversational naturalness, which sometimes means accepting slight imperfection in pronunciation if it makes the rhythm feel more human.
Growing pains in production environments
The community is actively working through implementation challenges. Users report that generated audio sometimes adds unwanted interjections and occasionally swallows words. Streaming audio generation produces noise in some configurations. When teams try to accelerate inference using vLLM, they encounter degraded speech quality.
These reflect ambitious goals meeting implementation reality. Controlling prosody at the level ChatTTS attempts means managing more variables than traditional TTS, and that complexity surfaces as rough edges in a project still finding its stride. The issues are well-documented and actively discussed—the community is validating the approach under real-world conditions.
The open-source TTS landscape
ChatTTS joins a group of open-source alternatives tackling similar problems from different angles. XTTS-v2, Bark, StyleTTS2, Piper TTS, and MeloTTS each make different tradeoffs between quality, speed, and controllability. Bark emphasizes environmental sounds, StyleTTS2 focuses on style transfer, Piper prioritizes low-latency inference.
What set ChatTTS apart in early HackerNews discussions was the prosody quality in dialogue contexts—the sense that two AI voices could sound like they were actually talking to each other, not reading adjacent scripts. That's a narrow but meaningful achievement in a field where improvements compound quickly.
When prosody control justifies the tradeoffs
For teams building conversational AI, the decision point is whether prosody naturalness matters enough to accept current limitations around unwanted interjections and streaming stability. If you're generating audiobook narration, traditional models with cleaner output probably make more sense. If you're prototyping a therapy chatbot where emotional tone carries meaning, or an accessibility tool where conversational rhythm aids comprehension, ChatTTS's approach deserves evaluation.
The 38,000+ GitHub stars suggest the community sees value in this direction, even while actively documenting what still needs work. Open-source progress rarely looks like polished releases—it looks like ambitious attempts with visible rough edges and transparent issue tracking.
2noise/ChatTTS
A generative speech model for daily dialogue.