🔥36氪•Stalecollected in 30m
Alibaba Launches 3 Mid-Size Qwen3.5 Models
💡Mid-size open LLMs beat GPT-4o mini, inference at $0.03/MTok equiv
⚡ 30-Second TL;DR
What Changed
Open-sourced Qwen3.5-35B-A3B, Qwen3.5-122B-A10B, Qwen3.5-27B
Why It Matters
Offers high-performance open models at low cost, enabling devs to rival proprietary LLMs without high inference expenses. Boosts accessible AI for global builders.
What To Do Next
Test Qwen3.5-Flash on AliCloud百炼 for 0.2 yuan/MTok inference today.
Who should care:Developers & AI Engineers
🧠 Deep Insight
Web-grounded analysis with 5 cited sources.
🔑 Enhanced Key Takeaways
- •Qwen3.5 series introduces native multimodal capabilities, unifying text, vision, and UI interaction in a single model trained on images, UI screenshots, and structured content for tasks like visual question answering and pixel-level grounding[1][2].
- •Qwen3.5-397B-A17B, recently open-sourced prior to mid-size models, ranks #3 on Artificial Analysis Intelligence Index with 45 score, behind GLM-5 (50) and Kimi K2.5 (47), featuring 262K context window and 17B active MoE parameters[2].
- •Qwen3.5-Plus hosted variant offers 1M token context window, 'Auto' mode with adaptive thinking and tools like search and code interpreter, outperforming Qwen3-Max in multimodal tasks at lower cost[1][3][4].
📊 Competitor Analysis▸ Show
| Feature/Model | Qwen3.5-397B-A17B | GLM-5 | Kimi K2.5 |
|---|---|---|---|
| Intelligence Index Score | 45 (#3 open weights) | 50 | 47 |
| Active Parameters (MoE) | 17B | 32B | 32B |
| Output Tokens (Index) | ~86M | 110M | 89M |
| Native Vision | Yes (image/video) | Not specified | Not specified |
🛠️ Technical Deep Dive
- •Qwen3.5-397B-A17B: 397B total parameters, 17B active (MoE architecture), 262K token context window, supports reasoning and non-reasoning modes in one model, native image/video input[2].
- •Qwen3.5-Plus: 1M token context window (extended), multimodal (text/image/video), includes 'Thinking', 'Fast', and 'Auto' modes with built-in tools (search, code interpreter), Max CoT up to 81,920 tokens[1][3][4].
- •Qwen3.5-397B-A17B hardware: FP16/BF16 requires ~800GB VRAM; 4-bit quantized ~220GB unified memory (runnable on Mac Studio/Pro M-series Ultra or multi-GPU setups like 3x A100 80GB)[1].
- •Unified training on text, images, UI screenshots, structured content; 19x faster decoding on 256K long-context vs. Qwen3-Max, 8.6x faster standard workflows without intelligence loss[1].
🔮 Future ImplicationsAI analysis grounded in cited sources
Qwen3.5 mid-size models will accelerate open-source MoE adoption in mid-tier deployments
Alibaba Cloud pricing at 0.2 yuan/million tokens will capture 20% more enterprise API market share
⏳ Timeline
2025-09
Qwen3 series released, establishing baseline with separate text and vision models
2025-09-11
Qwen3 initial release on GitHub
2025-12-01
Qwen-Plus (Qwen3 series) snapshot released with 32K context and batch pricing
2026-01-23
Qwen3-Max snapshot with integrated thinking modes and tool support (search, code interpreter)
2026-02-15
Qwen3.5-Plus snapshot released with 65K context, thinking by default, and multimodal enhancements
2026-02-16
Qwen3.5 series launched on GitHub with first 397B-A17B MoE model open-sourced
📎 Sources (5)
Factual claims are grounded in the sources below. Forward-looking analysis is AI-generated interpretation.
📰
Weekly AI Recap
Read this week's curated digest of top AI events →
👉Related Updates
AI-curated news aggregator. All content rights belong to original publishers.
Original source: 36氪 ↗