ViqusViqus
Navigate
Company
Blog
About Us
Contact
System Status
Enter Viqus Hub

Xiaomi Unleashes 1T Parameter MoE Model, Achieving Frontier Performance on Code and Agents.

MiMo-V2.5-Pro Mixture-of-Experts LLM Open-source MIT license LocalLLaMA
April 29, 2026
Source: AIModels.fyi
Viqus Verdict Logo Viqus Verdict Logo 8
Architectural Leap, Open Source Power Play
Media Hype 6/10
Real Impact 8/10

Article Summary

Xiaomi has quietly released MiMo-V2.5-Pro, a sophisticated Mixture-of-Experts (MoE) model, under the permissive MIT license. With 1.02 trillion total parameters (42 billion active per token), the model has benchmarked highly, exceeding current frontier models like Opus 4.6 in coding reasoning and agentic work. Technically, the model features a hybrid attention architecture, using sliding-window attention for efficiency and a multi-token prediction (MTP) module that significantly boosts inference speed. These innovations allow for a massive 1 million-token context window while maintaining speed and managing cache size effectively.

Key Points

  • MiMo-V2.5-Pro is an advanced 1.02T parameter MoE model released under the permissive MIT license.
  • The model shows state-of-the-art performance, particularly in coding reasoning and complex agentic decision-making, rivaling leading commercial benchmarks.
  • Key architectural enhancements include a hybrid attention mechanism (saving cache) and native multi-token prediction (tripling inference speed).

Why It Matters

This is not just another parameter announcement; the open release of a high-performance, architecturally optimized trillion-parameter MoE model is a significant development. The MIT license removes vendor lock-in concerns and accelerates community adoption for local, enterprise use. The specific focus on superior coding and agentic capabilities addresses two major pain points in current enterprise AI adoption. While demanding significant hardware resources, its open nature forces the competitive landscape to react, potentially driving down enterprise API costs and accelerating the democratization of top-tier LLM power.

You might also be interested in