MiniMax M2.5 Claims to Match Top AI Models at 5% the Cost
Chinese AI firm MiniMax releases M2.5, an open-source coding model claiming performance comparable to Claude and GPT-4 at dramatically lower prices.
Written by AI. Samira Okonkwo-Barnes
February 13, 2026

Photo: WorldofAI / YouTube
Chinese AI company MiniMax released its M2.5 model this week. The benchmark claims deserve a close look: 80.2% on SWE-Bench Verified, performance on par with Anthropic's Claude Opus and OpenAI's GPT-4, and pricing at $0.30 per million input tokens. That's roughly 20 times cheaper than top proprietary models.
The numbers are striking enough that WorldofAI's video demo has spread widely among developers. But benchmark scores and real-world value are different things. That's especially true for an open-source model from a company without the name recognition of Anthropic or OpenAI.
What the Benchmarks Actually Show
MiniMax's published benchmarks put M2.5 at 80.2% on SWE-Bench Verified. That's a coding test that checks whether models can fix real GitHub issues. For context, that's in the same range as Claude 3.5 Sonnet and GPT-4. The model also hits 76.3% on BrowserComp and 76.8% on agentic tool calling tasks.
These are claims, not third-party results. SWE-Bench scores can be gamed. Models can be trained to ace the benchmark without handling novel coding tasks well. The video creator tested M2.5 on several front-end tasks. It built a macOS-style browser interface, a Minecraft clone, and a Bloomberg-style investment portal.
"This is an open-source model, guys, that's going toe-to-toe with proprietary giants, whether that's models from Anthropic, Google, or even OpenAI," the video notes. The demos show solid UI work with working parts. But the creator also admits limits: "This is where I found the first error with this generation."
The error rate matters. One failed output in a demo tells you nothing about how reliable the model is at scale.
The Economics Deserve Attention
Pricing is where M2.5 gets genuinely interesting from a policy angle. At $0.30 per million input tokens and $1.20 per million output tokens, the model undercuts top options by a factor of 10 or more. MiniMax offers free access during a launch period. Paid plans start at about $0.30 per hour at 50 tokens per second.
This pricing creates real market pressure. If M2.5's claims hold up in outside tests, the cost gap forces a question: what exactly are enterprise buyers paying for with premium models?
The usual answers -- reliability, support, compliance tools, steady uptime -- still apply. But the gap between "good enough for most jobs" and "top-tier performance" may be thinner than premium prices suggest.
The 204.8K context window is worth noting. Most apps don't need contexts that large. But for niche tasks -- parsing long codebases, reviewing thick docs -- it removes a limit that used to set pricey models apart from cheaper ones.
The Open-Source Question
MiniMax calls M2.5 "open-source." But the license terms and weight access remain vague in the company's public docs. This matters a lot. "Open-source" in AI has become marketing speak. It can mean anything from "weights on a free license" to "you can use our API."
Truly open-source models let you inspect, change, and self-host them. They let researchers check training data sources, test for bias, and understand how the model works at a deep level. If M2.5 is open-source in that real sense, it adds something meaningful. If it just means cheap API access, that's useful but a very different thing.
The video shows M2.5 through Kilo Code and OpenCode, both autonomous coding agents that plug into VS Code. This pattern is becoming standard. Models now compete not just on raw power but on how well they fit into developer workflows.
What the Demos Don't Tell Us
The video's demos focus almost entirely on front-end web builds. That looks impressive but covers a narrow slice of coding. Front-end generation is mostly pattern matching over well-known frameworks. The hard problems -- debugging tangled state, tuning database queries, cleaning up legacy code -- don't show up in these demos.
The video creator notes: "You can run it for about $1 an hour at 100 tokens per second, which is just incredible, or roughly 30 cents an hour at 50 tokens per second, which finally makes the idea of intelligence too cheap to meter, which is truly realistic."
The phrase "intelligence too cheap to meter" echoes Lewis Strauss's famous 1954 claim about nuclear power. It's worth recalling how that turned out. Cheap compute still needs infrastructure. It still piles up costs at scale. It still hits reliability walls that only show up in production.
The Verification Gap
What's missing from this launch is outside testing. MiniMax is a Chinese company going up against American firms. The geopolitical angle adds tension around trust, data handling, and long-term access.
For enterprise buyers, these questions aren't minor. Using an AI model means sending potentially sensitive code and data through its systems. MiniMax's pricing shows aggressive market-share tactics. But whether that pricing lasts -- and where the company is headed -- stays unclear.
The model's claims need testing by independent researchers. Ideally, they'd use a broad mix of coding tasks beyond web UI work. Until that happens, M2.5 sits between interesting launch and proven option.
Market Pressure Is the Story
Whether or not M2.5 meets its benchmark claims, the bigger pattern is clear. The gap between proprietary frontier models and open alternatives keeps shrinking. Anthropic, OpenAI, and Google still lead in reliability, ecosystem, and cutting-edge power. But those edges cost a lot more. The premium gets harder to justify as open models close in on similar results.
This pricing pressure will likely speed up. MiniMax isn't alone in this game. Meta's Llama models, Mistral's lineup, and various Chinese AI labs all chase the same target: good-enough results at far lower cost.
For regulation, this means that arguments for limiting open-source AI on safety grounds must account for this dynamic. You can't restrict American open-source work while Chinese firms freely release strong models at bargain prices. The policy lever exists in theory. The enforcement mechanism grows more fictional by the day.
For developers weighing M2.5, the math is simple: test it on your own use case. Benchmarks give a signal, but your codebase has its own quirks. The free launch period makes trying it low-risk. Just know that promo pricing isn't production pricing. And demo performance often doesn't match production reality.
Samira Okonkwo-Barnes is Buzzrag's tech policy and regulation correspondent.
Watch the Original Video
MiniMax M2.5 IS INSANE! Best Opensource Coding Model! Beats Opus 4.6 and 20x Cheaper! (Fully Tested)
WorldofAI
10m 29sAbout This Source
WorldofAI
WorldofAI is an engaging YouTube channel that has swiftly captured the attention of AI enthusiasts, boasting 182,000 subscribers since its inception in October 2025. The channel is dedicated to showcasing the creative and practical applications of Artificial Intelligence in everyday tasks, offering viewers a rich collection of tips, tricks, and guides to enhance their daily and professional lives through AI.
Read full source profileMore Like This
Spec-Driven Development Tools Promise to Fix AI Coding
Tracer's Epic Mode tackles 'vibe coding' with structured specifications. But can better documentation really solve AI development's consistency problems?
Open-Source AI Agents Get Context Memory Via Airweave
Airweave turns workplace apps into searchable knowledge layers for AI agents, addressing the context problem that causes hallucinations and failures.
BMAD V6 Launches AI Development Platform Without Guardrails
BMAD V6 transforms AI coding into a modular platform, promising enterprise customization while raising questions about accountability and safety.
Cline CLI 2.0: Open-Source AI Coding Tool Goes Terminal
Cline CLI 2.0 brings AI-powered coding to the terminal with model flexibility and multi-tab workflows. But open-source AI tools raise questions.