←back to Blog

StepFun AI Releases Step-Audio 2 Mini: An Open-Source 8B Speech-to-Speech AI Model that Surpasses GPT-4o-Audio

StepFun AI Releases Step-Audio 2 Mini: An Open-Source 8B Speech-to-Speech AI Model that Surpasses GPT-4o-Audio

The StepFun AI team has released Step-Audio 2 Mini, an 8B parameter speech-to-speech large audio language model (LALM) that delivers expressive, grounded, and real-time audio interaction. Released under the Apache 2.0 license, this open-source model achieves state-of-the-art performance across speech recognition, audio understanding, and speech conversation benchmarks, surpassing commercial systems such as GPT-4o-Audio.

Target Audience Analysis

The primary audience for Step-Audio 2 Mini includes:

  • Developers looking for state-of-the-art speech technology for integration into applications.
  • Researchers aiming to advance the field of natural language processing (NLP) and machine learning (ML).
  • Business leaders in technology and communication sectors seeking innovative solutions to enhance user interaction.

Pain Points

Common challenges faced by the target audience include:

  • Difficulty in achieving high accuracy in speech recognition across diverse languages and dialects.
  • The need for seamless integration of audio and text processing within applications.
  • Challenges in creating emotionally aware conversational agents that can convey nuanced human interaction.

Goals

The audience’s goals likely involve:

  • Implementing advanced speech technologies that improve user experience and accessibility.
  • Exploring open-source solutions that allow for customization and innovation.
  • Staying ahead in competitive markets by leveraging cutting-edge AI advancements.

Key Features of Step-Audio 2 Mini

Unified Audio–Text Tokenization

Step-Audio 2 integrates Multimodal Discrete Token Modeling, allowing:

  • Seamless reasoning across text and audio.
  • On-the-fly voice style switching during inference.
  • Consistency in semantic, prosodic, and emotional outputs.

Expressive and Emotion-Aware Generation

This model interprets paralinguistic features such as pitch, rhythm, emotion, timbre, and style. Accordingly, Step-Audio 2 Mini achieves 83.1% accuracy on benchmarks, significantly outperforming GPT-4o Audio at 43.5%.

Retrieval-Augmented Speech Generation

Step-Audio 2 incorporates multimodal RAG (Retrieval-Augmented Generation), featuring:

  • Web search integration for factual grounding.
  • Audio search, enabling voice timbre/style imitation during inference.

Tool Calling and Multimodal Reasoning

The model supports tool invocation, achieving accuracy in tool selection that matches textual LLMs, while excelling in audio search tool calls, a feature absent in text-only LLMs.

Training and Data Scale

The model was trained on a 1.356T tokens text and audio corpus and over 8M+ hours of real and synthetic audio, featuring approximately 50K diverse voices across various languages and dialects.

Performance Benchmarks

In Automatic Speech Recognition (ASR), Step-Audio 2 achieves:

  • English: Average WER 3.14% (better than GPT-4o Transcribe at 4.5%).
  • Chinese: Average CER 3.08% (significantly lower than GPT-4o and Qwen-Omni).

For Audio Understanding (MMAU), it scores an average of 78.0, outperforming competitors. In Speech Translation (CoVoST 2), it achieved BLEU 39.26, the highest among models.

Conclusion

Step-Audio 2 Mini makes advanced multimodal speech intelligence accessible to the developer and research communities. By combining Qwen2-Audio’s reasoning capacity with CosyVoice’s tokenization pipeline, StepFun delivers one of the most capable open audio LLMs.

Further Exploration

Check out the model on Hugging Face. For more resources, visit our GitHub Page for tutorials, codes, and notebooks. Follow us on Twitter and join our 100k+ ML SubReddit. Don’t forget to subscribe to our newsletter!