←back to Blog

Baidu Open Sources ERNIE 4.5: LLM Series Scaling from 0.3B to 424B Parameters

Analysis of Target Audience for Baidu Open Sources ERNIE 4.5

The target audience for Baidu’s ERNIE 4.5 comprises AI researchers, business leaders, software developers, and data scientists interested in natural language processing (NLP) and machine learning (ML). This audience seeks advanced models capable of enhancing language understanding and generation capabilities across various applications.

Pain Points

  • Need for robust NLP solutions that can handle both Chinese and multilingual tasks efficiently.
  • Lack of access to powerful language models that support large-scale applications.
  • Challenges in integrating advanced AI models into existing systems.

Goals

  • To leverage state-of-the-art language models for improved business operations and user engagement.
  • To contribute to and benefit from the open-source AI community.
  • To experiment with diverse model architectures for specific use cases.

Interests

  • Latest developments in AI and NLP technologies.
  • Open-source projects and collaboration opportunities.
  • Performance benchmarks and model comparisons.

Communication Preferences

This audience prefers clear, concise, and technical communication, often seeking peer-reviewed data and practical applications over marketing language. They value detailed documentation and hands-on resources for model implementation.

Baidu Open Sources ERNIE 4.5: Technical Overview

Baidu has officially open-sourced its latest ERNIE 4.5 series, which includes a diverse family of foundation models designed for enhanced language comprehension, reasoning, and generation. The release features ten model variants, ranging from compact 0.3B dense models to substantial Mixture-of-Experts (MoE) architectures, with the largest variant reaching 424B parameters. These models are accessible to the global research and developer community via Hugging Face, promoting open experimentation and access to advanced Chinese and multilingual language technology.

ERNIE 4.5 Architecture

The ERNIE 4.5 series builds on prior iterations by introducing advanced model architectures, including both dense and sparsely activated MoE designs. Notably, the MoE variants efficiently scale parameter counts; for example, the ERNIE 4.5-MoE-3B and ERNIE 4.5-MoE-47B activate a limited number of experts per input token (typically 2 of 64 experts), which keeps the number of active parameters manageable while maintaining model expressivity and generalization capabilities.

These models utilize a combination of supervised fine-tuning (SFT), reinforcement learning with human feedback (RLHF), and contrastive alignment techniques. The training corpus encompasses 5.6 trillion tokens across diverse domains in both Chinese and English, employing Baidu’s proprietary multi-stage pretraining pipeline. The resulting models show high fidelity in instruction-following, multi-turn conversation, long-form generation, and reasoning benchmarks.

Model Variants and Open-Source Release

The ERNIE 4.5 release includes ten variants:

  • Dense Models: ERNIE 4.5-0.3B / 0.5B / 1.8B / 4B
  • MoE Models: ERNIE 4.5-MoE-3B / 4B / 6B / 15B / 47B / 424B total parameters (with varying active parameters)

The MoE-47B variant activates only 3B parameters during inference while maintaining a total of 47B. The 424B model—Baidu’s largest release—utilizes sparse activation strategies to make inference feasible and scalable. These models support FP16 and INT8 quantization for efficient deployment.

Performance Benchmarks

ERNIE 4.5 models demonstrate significant improvements across key Chinese and multilingual NLP tasks. According to the official technical report:

  • On CMMLU, ERNIE 4.5 surpasses earlier ERNIE versions, achieving state-of-the-art accuracy in Chinese language understanding.
  • On MMLU, the multilingual benchmark, ERNIE 4.5-47B shows competitive performance against leading LLMs like GPT-4 and Claude.
  • In long-form generation, ERNIE 4.5 receives higher coherence and factuality scores based on Baidu’s internal metrics.
  • Instruction-following tasks benefit from contrastive fine-tuning, leading to improved alignment with user intent and reduced hallucination rates compared to earlier ERNIE versions.

Applications and Deployment

ERNIE 4.5 models are optimized for various applications, including:

  • Chatbots and Assistants: With multilingual support and instruction-following alignment, these models are well-suited for AI assistants.
  • Search and Question Answering: High retrieval and generation fidelity enable integration with RAG pipelines.
  • Content Generation: Enhanced long-form text and knowledge-rich content generation is facilitated by improved factual grounding.
  • Code and Multimodal Extension: While currently focused on text, ERNIE 4.5 is compatible with multimodal extensions.

Supporting up to 128K context length in some variants, the ERNIE 4.5 family can be utilized in tasks requiring memory and reasoning across extensive documents or sessions.

Conclusion

The ERNIE 4.5 series marks a significant advancement in open-source AI development, offering a versatile set of models designed for scalable, multilingual, and instruction-aligned tasks. Baidu’s decision to release models ranging from lightweight 0.3B variants to a 424B-parameter MoE model highlights its commitment to inclusive and transparent AI research. With comprehensive documentation, open availability on Hugging Face, and support for efficient deployment, ERNIE 4.5 is set to accelerate global advancements in natural language understanding and generation.

Check out the Paper and Models on Hugging Face. All credit for this research goes to the researchers of this project. Also, feel free to follow us on Twitter and join our 100k+ ML SubReddit. Don’t forget to subscribe to our Newsletter.