🔥Stalecollected in 19m

Alibaba Qwen Launches Coding Plan

Alibaba Qwen Launches Coding Plan
PostLinkedIn
🔥Read original on 36氪

💡Qwen Coding Plan adds top models & tool compatibility—supercharge your AI coding workflow.

⚡ 30-Second TL;DR

What Changed

Launched Qwen Coding Plan for enhanced model support

Why It Matters

This expands Qwen's ecosystem for coding tasks, making it more accessible for developers integrating with popular tools. It strengthens Alibaba's position in AI coding against competitors.

What To Do Next

Test Qwen3-Coder-Next integration with ClaudeCode via the new Qwen Coding Plan.

Who should care:Developers & AI Engineers

🧠 Deep Insight

Web-grounded analysis with 8 cited sources.

🔑 Enhanced Key Takeaways

  • Alibaba's Qwen Coding Plan is a subscription service from Alibaba Cloud Model Studio offering up to 90,000 requests per month for AI coding tools like Qwen Code, Claude Code, Cline, and OpenClaw[2].
  • The plan supports new models including qwen3.5-plus, qwen3-max-2026-01-23, qwen3-coder-next, qwen3-coder-plus, and third-party models like glm-4.7 and kimi-k2.5[2].
  • Qwen3-Coder-Next is an open-weight model with 80B total parameters (3B activated via MoE), 256K context length, designed for local coding agents on consumer hardware[1][3][5].
  • Qwen3.5-Plus features hybrid reasoning (Thinking/Fast/Auto modes), up to 1M token context, 480B MoE architecture with 35B active parameters, and multimodal capabilities[1][4][7].
  • Qwen coding models excel in benchmarks: Qwen2.5-Coder at 88.4% HumanEval (beats GPT-4's 87.1%), 69.6% SWE-Bench, supporting 119 languages and local deployment on 32GB RAM[1].
📊 Competitor Analysis▸ Show
FeatureQwen Coding Plan / Qwen3-Coder-NextCompetitors (e.g., GPT-4o, Claude Sonnet 4.5)
BenchmarksHumanEval 88.4%, SWE-Bench 69.6%; Sonnet 4.5-level[1][3]GPT-4: HumanEval 87.1%; Claude competitive[1]
Parameters80B total, 3B active (MoE); 480B MoE (35B active)[1][3]Proprietary, larger active params typically[1]
Context Length256K-1M tokens[1][3][4]Varies, e.g., Claude up to 200K[3]
PricingFixed monthly fee, 90K requests; free open-source local[1][2]Per-token API; subscription tiers[1]
DeploymentLocal on 32-64GB RAM/consumer GPU; open-weight[1][3]Cloud-only proprietary[1]

🛠️ Technical Deep Dive

  • Qwen3-Coder-Next: 80B total parameters, 3B activated per inference using Mixture-of-Experts (MoE) with 512 experts (10 activated per token), Hybrid architecture (Gated DeltaNet + MoE + Gated Attention), 256K context, trained on large-scale executable tasks + RL, open-weight Apache-2.0 license[3][5][6].
  • Qwen3.5-Plus: 480B MoE with 35B active parameters, hybrid reasoning (Thinking mode with 1024 token budget, Fast, Auto with tools), 1M token context, 250K vocabulary, multi-token prediction (10-60% token savings), native multimodal (text/images/UI), 119 languages including programming[1][4][7].
  • Coding Plan Models: Includes qwen3-max-2026-01-23 (thinking enabled), supports third-party integration via Model Context Protocol (MCP) for agents[2].

🔮 Future ImplicationsAI analysis grounded in cited sources

Qwen Coding Plan democratizes advanced AI coding with affordable subscriptions, open-weight local models, and superior benchmarks, challenging proprietary leaders like OpenAI and Anthropic by enabling privacy-focused, cost-effective agentic coding on consumer hardware, potentially accelerating open-source adoption in enterprise development[1][2][3].

Timeline

2026-01
Release of qwen3-max-2026-01-23, flagship Qwen3 model with thinking capabilities[2]
2026-02
Launch of Qwen3.5 series including Qwen3.5-Plus and Qwen3-Coder-Next, alongside Coding Plan subscription[1][2][3][4]
📰

Weekly AI Recap

Read this week's curated digest of top AI events →

👉Related Updates

AI-curated news aggregator. All content rights belong to original publishers.
Original source: 36氪