๐Ÿ‡ญ๐Ÿ‡ฐStalecollected in 2m

Chinese Models Top Global AI Token Usage

Chinese Models Top Global AI Token Usage
PostLinkedIn
๐Ÿ‡ญ๐Ÿ‡ฐRead original on SCMP Technology

๐Ÿ’กChinese open-source LLMs surpass US models in global usageโ€”benchmark for your next project.

โšก 30-Second TL;DR

What Changed

MiniMax M2.5 tops OpenRouter token usage ranking

Why It Matters

Highlights growing competitiveness of Chinese AI firms in open-source space, offering cost-effective high-performance alternatives to US models. AI practitioners gain more options for scalable deployments via platforms like OpenRouter.

What To Do Next

Deploy MiniMax M2.5 via OpenRouter API to test its top-ranked token efficiency.

Who should care:Developers & AI Engineers

๐Ÿง  Deep Insight

Web-grounded analysis with 9 cited sources.

๐Ÿ”‘ Enhanced Key Takeaways

  • โ€ขChinese models captured 61% of total token volume on OpenRouter, with MiniMax M2.5 at 2.45 trillion tokens, Kimi K2.5 at 1.21 trillion, and Zhipu's GLM-5 at 780 billion.[2][3]
  • โ€ขMiniMax M2.5 launched on February 13, 2026, as the worldโ€™s first production-grade model natively designed for agent scenarios, achieving 3.07 trillion tokens in its first seven days and a 197% week-over-week surge.[2]
  • โ€ขMoonshot AI's Kimi K2.5 generated revenue in under 20 days post-launch exceeding its entire 2025 total, with overseas revenue surpassing domestic for the first time, fueled by global paid subscribers and API usage.[2]
  • โ€ขPricing advantage is key: MiniMax M2.5 and GLM-5 at $0.30 per million input tokens vs. Claude Opus 4.6 at $5.00, making them 16.7 times cheaper.[2]
๐Ÿ“Š Competitor Analysisโ–ธ Show
ModelParametersContext WindowKey BenchmarksPricing (Input/Output per 1M tokens)
MiniMax M2.5230B205KSWE-bench Verified 80.2, HumanEval 89.6$0.30 / ?
Kimi K2.51T total (32B active MoE)262KHumanEval 99.0, MMLU 92.0, MATH-500 98.0?
GLM-5??Top coding usage$0.30 / ?
Claude Opus 4.6??Lower token usage$5.00 / ?

๐Ÿ› ๏ธ Technical Deep Dive

  • โ€ขMiniMax M2.5: 230B parameters, 205K context window; excels in real-world software engineering with SWE-bench Verified 80.2 (highest), Multi-SWE-Bench 51.3, BrowseComp 76.3; trained for agent scenarios, office tools (Word, Excel, PowerPoint), context switching, and token-efficient planning.[1][4]
  • โ€ขKimi K2.5: 1T total parameters (32B active per token, MoE architecture), 262K context window; native multimodal with visual coding and self-directed agent swarm; standout benchmarks include HumanEval 99.0, MMLU 92.0, MMLU-Pro 87.1, LiveCodeBench 85.0, AIME 2025 96.1, GPQA Diamond 87.6, MATH-500 98.0, Chatbot Arena 1447, IFEval 94.0; continued pretraining on 15T mixed visual/text tokens.[1][4]

๐Ÿ”ฎ Future ImplicationsAI analysis grounded in cited sources

Chinese models will exceed 70% OpenRouter share by Q2 2026
Driven by pricing 16.7x lower than US leaders, superior coding/agent benchmarks, and surging global demand as shown in 61% share and revenue records.[2]
Agentic workflows will define LLM market leadership
Top models like M2.5 prioritize native agent design, SWE-bench dominance, and tool-calling, shifting competition from general reasoning to practical automation.[2][4]
Moonshot AI valuation will double post-K2.5 revenue surge
Already China's fastest decacorn with 20-day revenue topping 2025 total and overseas surpassing domestic, signaling hyper-growth trajectory.[2]

โณ Timeline

2026-02-13
MiniMax launches M2.5, world's first production-grade agent-native model
2026-02
Moonshot AI releases Kimi K2.5, achieving record revenue surge
2026-02-24
OpenRouter data shows Chinese models at 61% token share, MiniMax M2.5 #1
๐Ÿ“ฐ

Weekly AI Recap

Read this week's curated digest of top AI events โ†’

๐Ÿ‘‰Related Updates

AI-curated news aggregator. All content rights belong to original publishers.
Original source: SCMP Technology โ†—