๐Ÿ‡ญ๐Ÿ‡ฐStalecollected in 1m

Alibaba Qwen3.5-Max Tops China, Trails US

Alibaba Qwen3.5-Max Tops China, Trails US
PostLinkedIn
๐Ÿ‡ญ๐Ÿ‡ฐRead original on SCMP Technology

๐Ÿ’กChina's top LLM beats peers on benchmarks, closing US gap

โšก 30-Second TL;DR

What Changed

Qwen3.5-Max-Preview tops Chinese AI models on Arena rankings

Why It Matters

Strengthens China's domestic AI ecosystem, offering practitioners a high-performing alternative to US models. May accelerate competition and innovation in multimodal LLMs.

What To Do Next

Test Qwen3.5-Max-Preview on Arena to benchmark against Claude and GPT models.

Who should care:Researchers & Academics

๐Ÿง  Deep Insight

Web-grounded analysis with 7 cited sources.

๐Ÿ”‘ Enhanced Key Takeaways

  • โ€ขThe Qwen 3.5 family introduces a novel 'Hybrid Mixture-of-Experts' architecture utilizing Gated DeltaNet (linear attention), which enables a 1-million token context window while delivering up to 19x higher decoding throughput than the previous Qwen 3 generation.
  • โ€ขAlibaba has expanded linguistic support to 201 languages and dialects, utilizing a massive 250,000-token vocabulary that improves encoding efficiency by up to 60% for non-English scripts compared to the 150,000-token limit in Qwen 3.
  • โ€ขThe release follows a significant leadership exodus in early 2026, including the departure of technical lead Lin Junyang (Justin Lin) and head of post-training Yu Bowen, sparking industry debate over Alibaba's long-term commitment to its open-source strategy.
  • โ€ขQwen 3.5-Max-Preview features a dual-mode 'Thinking' vs. 'Fast' inference capability, where the model can engage in internal chain-of-thought reasoning (via <think> tags) to match US rivals in complex logic while maintaining a low-latency mode for routine tasks.
๐Ÿ“Š Competitor Analysisโ–ธ Show
FeatureQwen 3.5-Max-PreviewGemini 3.1 ProClaude 4.6 OpusGPT-5.4
Arena Elo~1451150515031485
Context Window1M (Hosted) / 262K (Native)2M+200K128K
Architecture397B MoE (17B Active)Proprietary MoEProprietaryProprietary
LicenseApache 2.0 (Open-Weight)ProprietaryProprietaryProprietary
Multilingual201 Languages150+ Languages95+ Languages100+ Languages
Pricing (per 1M)~$0.10 (Est. API)$1.25 (Input)$3.00 (Input)$2.50 (Input)

๐Ÿ› ๏ธ Technical Deep Dive

Detailed technical specifications for the Qwen 3.5-397B-A17B model:

  • Parameter Count: 397 billion total parameters with a sparse Mixture-of-Experts (MoE) routing that activates only 17 billion parameters per token.
  • Attention Mechanism: A hybrid layout consisting of 60 layers where 15 groups of 3 'Gated DeltaNet' (linear attention) layers are interleaved with 1 'Gated Attention' layer to optimize memory usage for long-context sequences.
  • Multimodal Integration: Native 'early-fusion' vision-language architecture where text and visual tokens are processed within the same transformer backbone rather than using a separate adapter.
  • Training Scale: Pre-trained on an estimated 36+ trillion tokens with a heavy emphasis on synthetic 'agentic' data and reinforcement learning (RL) scaled across million-agent environments.
  • Inference Optimizations: Native support for Multi-Token Prediction (MTP) and SGLang/vLLM acceleration engines, achieving near-100% multimodal training efficiency.

๐Ÿ”ฎ Future ImplicationsAI analysis grounded in cited sources

Alibaba will pivot toward a 'Cloud-First' proprietary model tier.
The increasing performance gap between the open-weight 397B model and the closed-source 'Plus' and 'Max-Preview' versions suggests Alibaba is prioritizing its Model Studio ecosystem over pure open-source parity.
Qwen will become the dominant foundation for non-English AI agents.
With support for 201 languages and superior performance on regional benchmarks like C-Eval, it is positioned as the primary alternative to US models in Southeast Asia and the Middle East.

โณ Timeline

2023-04
Tongyi Qianwen (Qwen) Beta Launch
2024-06
Qwen 2 Series Released with 72B Flagship
2024-09
Qwen 2.5 Launch with Enhanced Reasoning
2025-04
Qwen 3 Family Debut (Apache 2.0 License)
2026-02
Qwen 3.5 and 397B MoE Open-Weight Release
2026-03
Qwen 3.5-Max-Preview Deployed on LMSYS Arena

๐Ÿ“Ž Sources (7)

Factual claims are grounded in the sources below. Forward-looking analysis is AI-generated interpretation.

  1. vertexaisearch.cloud.google.com โ€” Auziyqgx 5rr4y0wjivjfqy4rlynpef Goo in Ww5bjku6k7tjrhfzbamegop Lzw Pcprivqgeaumnzcxesu8leg3qutabz89x7qhedjmdgwuteyckgxr5cvvsf9w=
  2. vertexaisearch.cloud.google.com โ€” Auziyqgtzddovplndy05twxbslji Wa4qqtrv2ev9cvj0jowqyu9 Die68 Pgcparg0zqyqfcjewjuzastybzbs Kl1b9fimkgwyrvskf2yie4v0z6xroauwqdzvqhwlciroozg=
  3. vertexaisearch.cloud.google.com โ€” Auziyqeybs Zfcednytpwugnmdxbmmceu92sl6hg5lxa2kifdro8v5ztkt17m9r Dpgfcnik5jynthpcdf79swna2aknlml Vy17doftkud4prqed05zclcunalw9hu9
  4. vertexaisearch.cloud.google.com โ€” Auziyqf1seglzawbrhyftekrmismojq0vk3 Kd2wzvlheog8wok83u4pchyisbdhfcrj7vtlbqnemwc0 Jrrtgjj Mu0k5p6j7n6icjzff2adjwsdzahvodkwrlv8pkkugiwfrga Jy2a0ksba2bhvgsgxucgddfdkcupvtfn3ylt Afrkv40pxeqjhtr666rj9czyqrij3l Rrrkmv6mbz5gkoyqgo1pdhnofkztncqoy1pogsy4wkbl Hhdewj6f2bym2qwwpopk 4hv1dneq1cxdk
  5. vertexaisearch.cloud.google.com โ€” Auziyqfz778jcjk4nowmmdgphwj0p7kwds9rrwo4knrk9lnacxajq8ad Oi0swrwhhk5tvmrb6ctgs0rdnllsm6vebnfd1ve4eaudugjlrhn5tt6qmywoi7sldu Fopf6irikbt0ybhj4 Dwy0pv1krguteq0piepmfc 31ldjyvnq Ll1vjfieaiqtnnlhb3qk7vfxs894d4m4fdrcrfzmr3igxr80gxxlm0lpzssxa4 Kznfqgrkw M Ozpxfncyaqhfo=
  6. vertexaisearch.cloud.google.com โ€” Auziyqfowhpdfuhcpna3usbpgmabdjybobqpyd4pfu7idr Hsmtfdthiwyuf3mykxqivle86cblr1a44lfcpl7mtiuwqfjmhjxygp8f4mmxhcrqwkhja36 Clopielmrpc57da8cztk=
  7. vertexaisearch.cloud.google.com โ€” Auziyqf1qyy5fbppohuqxue5rzoy 4sz48lbuyrbcl8haypfg4uao9suewrrscovmiz Vpcqd0aphtfxefvmx 2dzfrodqqnrqd4q16h1chghzggymuxtzn7panukbgpmtslmmvqsr4miedsgu1x Ksmpwz8oudjoa==
๐Ÿ“ฐ

Weekly AI Recap

Read this week's curated digest of top AI events โ†’

๐Ÿ‘‰Related Updates

AI-curated news aggregator. All content rights belong to original publishers.
Original source: SCMP Technology โ†—