←back to Blog

Kyutai Releases 2B Parameter Streaming Text-to-Speech TTS with 220ms Latency and 2.5M Hours of Training

Kyutai Releases 2B Parameter Streaming Text-to-Speech TTS with 220ms Latency and 2.5M Hours of Training

Understanding the Target Audience

The target audience for Kyutai’s release includes:

  • AI researchers focused on speech synthesis technologies
  • Developers and engineers building voice-enabled applications
  • Businesses seeking scalable and efficient TTS solutions

Their pain points often revolve around:

  • High latency in existing TTS systems
  • Limited multilingual support
  • Access to open-source tools for experimentation and development

Goals of the audience include:

  • Implementing real-time TTS in applications
  • Enhancing user experience through responsive voice interfaces
  • Achieving efficiencies in AI deployment for better cost management

In terms of communication preferences, this audience typically favors:

  • Technical documentation and specifications
  • Community forums and collaborative platforms
  • Detailed case studies demonstrating practical applications

Product Overview

Kyutai, an open AI research lab, has introduced an advanced streaming Text-to-Speech (TTS) model with ~2 billion parameters. This model features ultra-low latency audio generation at 220 milliseconds while maintaining high fidelity. It is trained on 2.5 million hours of audio and licensed under the CC-BY-4.0 license, promoting openness and reproducibility.

Performance Highlights

The model supports a maximum of 32 concurrent users on a single NVIDIA L40 GPU while ensuring latency remains under 350 milliseconds. For individual users, it achieves latency as low as 220 milliseconds, enabling near real-time applications such as:

  • Conversational agents
  • Voice assistants
  • Live narration systems

This performance is made possible by Kyutai’s innovative Delayed Streams Modeling approach, which generates speech incrementally as the text is received, contrasting with traditional autoregressive models that experience response delays.

Key Technical Metrics

Here are the crucial specifications of the TTS model:

  • Model size: ~2B parameters
  • Training data: 2.5 million hours of speech
  • Latency: 220 ms for a single user, <350 ms for up to 32 users on one L40 GPU
  • Language support: English and French
  • License: CC-BY-4.0 (open source)

Delayed Streams Modeling Explained

Kyutai’s Delayed Streams Modeling technique allows speech synthesis to commence before the complete input text is available, balancing prediction quality with response speed for high-throughput streaming TTS. This method maintains temporal coherence, achieving faster-than-real-time synthesis.

The codebase and training recipe for this architecture are available on Kyutai’s GitHub repository, supporting reproducibility and community contributions.

Model Availability and Open Research Commitment

Kyutai has made the model weights and inference scripts available on Hugging Face, facilitating access for researchers and developers. The CC-BY-4.0 license allows for the unrestricted adaptation and integration of the model, as long as proper attribution is provided.

This release supports both batch and streaming inference, making it suitable for applications such as:

  • Voice cloning
  • Real-time chatbots
  • Accessibility tools

Kyutai’s multilingual TTS capabilities set a robust foundation for diverse applications.

Implications for Real-Time AI Applications

By reducing latency to approximately 200 ms, Kyutai’s model minimizes the delay between user intent and speech output. This is crucial for:

  • Conversational AI with human-like voice interfaces
  • Assistive tech such as screen readers and voice feedback systems
  • Media production that requires rapid voiceovers
  • Edge devices designed for low-power environments

The ability to support 32 concurrent users on a single GPU without sacrificing quality makes it an attractive option for efficiently scaling speech services in cloud environments.

Conclusion: Open, Fast, and Ready for Deployment

Kyutai’s streaming TTS release marks a significant advancement in speech AI. With superior synthesis quality, real-time latency, and generous licensing, it addresses vital needs for both researchers and product teams. Its reproducibility, multilingual support, and scalable performance offer a competitive alternative to proprietary solutions.