Text-to-Speech
A speech synthesis capability that converts written text into natural-sounding audio speech, using neural models trained on human speech recordings to generate expressive, voice-cloned, or custom-voiced audio.
Neural text-to-speech (TTS) models like Tacotron 2, FastSpeech, and their successors generate mel-spectrograms from phoneme sequences, then convert spectrograms to audio waveforms using vocoders (HiFi-GAN, WaveNet). Diffusion-based TTS models (Voicebox, VoiceBox 2) and flow matching approaches (Kokoro, Cartesia Sonic) produce state-of-the-art naturalness and emotion control.
Zero-shot voice cloning—reproducing a target voice from a short reference audio clip—has become reliable with models like XTTS, Parler-TTS, and Chatterbox (open-source). This capability powers personalized assistant voices, audiobook narration, and localization. Commercial ElevenLabs pioneered high-quality cloning; open-source alternatives now achieve comparable quality.
For AI agent pipelines, TTS enables voice interfaces without ElevenLabs dependency. Kokoro and OmniVoice provide MIT-licensed TTS with multiple voice styles suitable for production use. Latency is a key metric: streaming TTS models begin producing audio before completing synthesis, achieving first-byte latencies under 200ms.