Xiaomi stuns with new MiMo-V2-Professional LLM nearing GPT-5.2, Opus 4.6 efficiency at a fraction of the price
Chinese language electronics and automobile producer Xiaomi stunned the worldwide AI neighborhood at the moment with the release of MiMo-V2-Pro, a brand new 1-trillion parameter basis mannequin with benchmarks approaching these of U.S. AI giants OpenAI and Anthropic, however at round a seventh or sixth the price when accessed over proprietary API — and importantly, sending lower than 256,000 tokens-worth of data backwards and forwards.
Led by Fuli Luo, a veteran of the disruptive DeepSeek R1 challenge, the discharge represents what Luo characterizes as a "quiet ambush" on the worldwide frontier. Moreover, Luo acknowledged in an X post that the corporate does plan to open supply a mannequin variant from this newest launch, " when the fashions are secure sufficient to deserve it."
By specializing in the "motion area" of intelligence—shifting from code technology to the autonomous operation of digital "claws"—Xiaomi is trying to leapfrog the conversational paradigm solely.
Previous to this foray into frontier AI, Beijing-based Xiaomi established itself as a titan of "The Web of Issues" and client {hardware}.
Globally acknowledged because the world’s third-largest smartphone producer, Xiaomi spent the early 2020s executing a high-stakes entry into the automotive sector. Its electrical automobiles (EVs), such because the SU7 and the not too long ago launched YU7 SUV, have turned the corporate right into a vertically built-in powerhouse able to merging {hardware}, software program, and now, superior reasoning.
This pedigree in physical-world engineering informs MiMo-V2-Professional’s structure; it’s constructed to be the "mind" of complicated methods, whether or not these methods are managing international provide chains or navigating the intricate scaffolds of an autonomous coding agent.
Expertise: The structure of company
The central problem of the "Agent Period" is sustaining high-fidelity reasoning over huge spans of information with out incurring a prohibitive "intelligence tax" in latency or price. MiMo-V2-Professional addresses this via a sparse structure: whereas it homes 1T complete parameters, solely 42B are lively throughout any single ahead cross, making it roughly 3 times the scale of its predecessor, MiMo-V2-Flash.
The mannequin’s effectivity is rooted in an developed Hybrid Consideration mechanism. Normal transformers usually face a quadratic enhance in compute necessities as context grows; MiMo-V2-Professional makes use of a 7:1 hybrid ratio (elevated from 5:1 within the Flash model) to handle its huge 1M-token context window. This architectural selection permits the mannequin to keep up a deep "reminiscence" of long-running duties with out the efficiency degradation normally seen in frontier fashions.
The analogy: Consider the mannequin not as a pupil studying a guide page-by-page, however as an knowledgeable researcher in an unlimited library. The 7:1 ratio permits the mannequin to "skim" 85% of the information for context whereas making use of high-density consideration to the 15% most related to the duty at hand.
That is paired with a light-weight Multi-Token Prediction (MTP) layer, which permits the mannequin to anticipate and generate a number of tokens concurrently, drastically decreasing the latency required for the "considering" phases of agentic workflows. In accordance with Luo, these structural choices have been made months upfront, particularly to offer a "structural benefit" for the sudden pace at which the trade shifted towards brokers.
Product and benchmarking: A 3rd-party actuality verify
Xiaomi’s inside knowledge paints an image of a mannequin that excels in "real-world" duties over artificial benchmarks. On GDPval-AA, a benchmark measuring efficiency on agentic real-world work duties, MiMo-V2-Professional achieved an Elo of 1426, putting it forward of main Chinese language friends like GLM-5 (1406) and Kimi K2.5 (1283).
Whereas it nonetheless trails Western "max effort" fashions like Claude Sonnet 4.6 (1633) in uncooked Elo, it represents the best recorded efficiency for a Chinese language-origin mannequin on this class.
The third-party benchmarking group Artificial Analysis verified these claims, putting MiMo-V2-Professional at #10 on its international Intelligence Index with a rating of 49. This locations it in the identical tier as GPT-5.2 Codex and forward of Grok 4.20 Beta. These outcomes counsel that Xiaomi has efficiently constructed a mannequin able to the high-level reasoning required for engineering and manufacturing duties.
Key metrics from Synthetic Evaluation spotlight a major leap over the earlier open-weights model, MiMo-V2-Flash (which scored 41):
-
Hallucination charge: The Professional mannequin decreased hallucination charges to 30%, a pointy enchancment over the Flash mannequin’s 48%.
-
Omniscience index: It scored a +5, putting it forward of GLM-5 (+2) and Kimi K2.5 (-8).
-
Token effectivity: To run the complete Intelligence Index, MiMo-V2-Professional required solely 77M output tokens, considerably lower than GLM-5 (109M) or Kimi K2.5 (89M), indicating a extra concise and environment friendly reasoning course of.
Xiaomi’s personal charts additional emphasize its "Normal Agent" and "Coding Agent" capabilities. On ClawEval, a benchmark for agentic scaffolds, the mannequin scored 61.5, approaching the efficiency of Claude Opus 4.6 (66.3) and considerably outpacing GPT-5.2 (50.0). In coding-specific environments like Terminal-Bench 2.0, it achieved an 86.7, suggesting excessive reliability when executing instructions in a reside terminal atmosphere.
How enterprises ought to consider MiMo-V2-Professional for utilization
For the personas outlined in modern AI organizations—from Infrastructure to Safety—MiMo-V2-Professional represents a paradigm shift within the "Value-High quality" curve.
Infrastructure decision-makers will discover MiMo-V2-Professional a compelling candidate for the Pareto frontier of intelligence vs. price. Synthetic Evaluation reported that operating their index price solely $348 for MiMo-V2-Professional, in comparison with $2,304 for GPT-5.2 and $2,486 for Claude Opus 4.6.
For organizations managing GPU clusters or procurement, the power to entry top-10 international intelligence at roughly 1/seventh the price of Western incumbents is a robust incentive for production-scale testing.
Knowledge decision-makers can leverage the 1M context window for RAG-ready architectures, permitting them to feed total enterprise codebases or documentation units right into a single immediate with out the fragmentation required by smaller context fashions.
A methods/orchestration decision-maker ought to consider MiMo-V2-Professional as a main "mind" for multi-agent coordination. As a result of the mannequin is optimized for OpenClaw and Claude Code, it could possibly deal with long-horizon planning and exact device use with out the fixed human intervention that plagues earlier fashions.
Its excessive rating in GDPval-AA suggests it’s significantly well-suited for the workflow and orchestration layer wanted to scale AI throughout the enterprise. It permits for the creation of methods that may transfer past easy automation into complicated, multi-step downside fixing.
Nevertheless, safety decision-makers should train warning. The very "agentic" nature that makes the mannequin highly effective—its potential to make use of terminals and manipulate recordsdata—will increase the floor space for immediate injection and unauthorized mannequin entry.
Whereas its low hallucination charge (30%) is a defensive boon, the shortage of public weights (not like the Flash model) means inside safety groups can not carry out the deep "model-level" audits generally required for extremely delicate deployments. Any enterprise implementation have to be accompanied by strong monitoring and auditability protocols.
Pricing, availability, and the trail ahead
Xiaomi has priced MiMo-V2-Professional to dominate the developer market. The pricing is tiered primarily based on context utilization, with aggressive charges for caching to assist high-frequency reasoning duties.
-
MiMo-V2-Professional (as much as 256K): $1 per 1M enter tokens and $3 per 1M output tokens
-
MiMo-V2-Professional (256K-1M): $2 per 1M enter tokens and $6 per 1M output tokens
-
Cache learn: $0.20 per 1M tokens for the decrease tier and $0.40 for the upper tier
-
Cache write: Briefly free ($0)
Right here's the way it stacks as much as different main frontier fashions all over the world:
|
Mannequin |
Enter |
Output |
Whole Price |
Supply |
|
Grok 4.1 Quick |
$0.20 |
$0.50 |
$0.70 |
|
|
MiniMax M2.7 |
$0.30 |
$1.20 |
$1.50 |
|
|
Gemini 3 Flash |
$0.50 |
$3.00 |
$3.50 |
|
|
Kimi-K2.5 |
$0.60 |
$3.00 |
$3.60 |
|
|
MiMo-V2-Professional (≤256K) |
$1.00 |
$3.00 |
$4.00 |
|
|
GLM-5-Turbo |
$0.96 |
$3.20 |
$4.16 |
|
|
GLM-5 |
$1.00 |
$3.20 |
$4.20 |
|
|
Claude Haiku 4.5 |
$1.00 |
$5.00 |
$6.00 |
|
|
Qwen3-Max |
$1.20 |
$6.00 |
$7.20 |
|
|
Gemini 3 Professional |
$2.00 |
$12.00 |
$14.00 |
|
|
GPT-5.2 |
$1.75 |
$14.00 |
$15.75 |
|
|
GPT-5.4 |
$2.50 |
$15.00 |
$17.50 |
|
|
Claude Sonnet 4.5 |
$3.00 |
$15.00 |
$18.00 |
|
|
Claude Opus 4.6 |
$5.00 |
$25.00 |
$30.00 |
|
|
GPT-5.4 Professional |
$30.00 |
$180.00 |
$210.00 |
This aggressive positioning is designed to encourage the high-intensity software flows that outline the following technology of software program. The mannequin is at the moment obtainable by way of Xiaomi’s first-party API solely, with no present assist for picture or multimodal enter—a notable omission in an period of "Omni" fashions, although Xiaomi has teased a separate MiMo-V2-Omni for these wants.
The "Hunter Alpha" interval on OpenRouter proved that the market has a excessive urge for food for this particular mix of effectivity and reasoning. Fuli Luo’s philosophy—that analysis velocity is fueled by a "real love for the world you're constructing for"—has resulted in a mannequin that ranks 2nd in China and eighth worldwide on established intelligence indices.
Whether or not it stays a "quiet" ambush or turns into the inspiration for a world realignment of AI energy is dependent upon how rapidly builders undertake the "motion area" over the "chat window". For now, Xiaomi has moved the goalposts: the query is not simply "can it discuss?" however "can it act?"
Source link
latest video
latest pick
news via inbox
Nulla turp dis cursus. Integer liberos euismod pretium faucibua














