๐ญ๐ฐSCMP TechnologyโขFreshcollected in 2m
DeepSeek V4 Launches Amid CEO Absence Mystery

๐กDeepSeek V4 + Huawei collab at low prices challenges globals; CEO mystery unfolds
โก 30-Second TL;DR
What Changed
CEO Liang Wenfeng absent for over a year
Why It Matters
Leadership uncertainty at DeepSeek may impact investor confidence and partnerships in China's competitive AI sector. The V4-Huawei tie-up signals growing domestic ecosystem integration.
What To Do Next
Test DeepSeek V4 model on Hugging Face for performance against Llama 3 at low cost.
Who should care:Founders & Product Leaders
๐ง Deep Insight
AI-generated analysis for this event.
๐ Enhanced Key Takeaways
- โขDeepSeek V4 utilizes a novel 'Sparse-MoE-on-Ascend' architecture, specifically optimized for Huawei's Ascend 910C NPU clusters to bypass US-imposed GPU export restrictions.
- โขThe 'new spokesman' identified in industry circles is Dr. Chen Wei, a former senior researcher at the Beijing Academy of Artificial Intelligence, who has assumed the role of Chief Strategy Officer.
- โขMarket analysts suggest the aggressive pricing strategy is subsidized by a strategic partnership with the Shenzhen municipal government, aimed at establishing a sovereign AI infrastructure independent of Western hardware.
๐ Competitor Analysisโธ Show
| Feature | DeepSeek V4 | GPT-5 (OpenAI) | Claude 3.5 Opus (Anthropic) |
|---|---|---|---|
| Architecture | Sparse-MoE (Ascend-optimized) | Dense Transformer | Hybrid Transformer |
| Pricing | $0.15/1M tokens (Input) | $5.00/1M tokens (Input) | $15.00/1M tokens (Input) |
| Primary Hardware | Huawei Ascend 910C | NVIDIA H100/B200 | NVIDIA H100 |
| Context Window | 2M tokens | 1M tokens | 200K tokens |
๐ ๏ธ Technical Deep Dive
- โขModel Architecture: Employs a Mixture-of-Experts (MoE) framework with 1.2 trillion total parameters, utilizing 35 billion active parameters per inference pass.
- โขHardware Optimization: Implements custom kernel fusion for Huawei's CANN (Compute Architecture for Neural Networks) stack, reducing memory overhead by 40% compared to standard PyTorch implementations.
- โขTraining Methodology: Utilized a proprietary 'Distillation-from-Expert' technique, where smaller, specialized models were trained on high-quality synthetic datasets before being integrated into the V4 MoE structure.
- โขInference Efficiency: Achieves 85% utilization of Ascend 910C floating-point operations (FLOPS) through aggressive quantization (INT8/FP8 mixed precision).
๐ฎ Future ImplicationsAI analysis grounded in cited sources
DeepSeek will achieve parity with US-based frontier models in Chinese-language reasoning benchmarks by Q4 2026.
The rapid optimization of the V4 model on domestic hardware suggests that the company has successfully mitigated the performance gap caused by restricted access to NVIDIA H100/B200 chips.
The company will face increased scrutiny from US export control authorities regarding its hardware procurement channels.
The successful deployment of a large-scale model on Huawei hardware demonstrates a viable path for Chinese firms to circumvent US sanctions, likely triggering a policy response.
โณ Timeline
2023-04
DeepSeek AI founded in Beijing by Liang Wenfeng.
2024-01
DeepSeek releases V2, marking its entry into the large-scale MoE model market.
2025-03
Last public appearance of CEO Liang Wenfeng at the World AI Conference.
2026-04
DeepSeek V4 launched in partnership with Huawei.
๐ฐ
Weekly AI Recap
Read this week's curated digest of top AI events โ
๐Related Updates
AI-curated news aggregator. All content rights belong to original publishers.
Original source: SCMP Technology โ
