๐Ÿ‡ญ๐Ÿ‡ฐFreshcollected in 2m

DeepSeek V4 Launches Amid CEO Absence Mystery

DeepSeek V4 Launches Amid CEO Absence Mystery
PostLinkedIn
๐Ÿ‡ญ๐Ÿ‡ฐRead original on SCMP Technology

๐Ÿ’กDeepSeek V4 + Huawei collab at low prices challenges globals; CEO mystery unfolds

โšก 30-Second TL;DR

What Changed

CEO Liang Wenfeng absent for over a year

Why It Matters

Leadership uncertainty at DeepSeek may impact investor confidence and partnerships in China's competitive AI sector. The V4-Huawei tie-up signals growing domestic ecosystem integration.

What To Do Next

Test DeepSeek V4 model on Hugging Face for performance against Llama 3 at low cost.

Who should care:Founders & Product Leaders

๐Ÿง  Deep Insight

AI-generated analysis for this event.

๐Ÿ”‘ Enhanced Key Takeaways

  • โ€ขDeepSeek V4 utilizes a novel 'Sparse-MoE-on-Ascend' architecture, specifically optimized for Huawei's Ascend 910C NPU clusters to bypass US-imposed GPU export restrictions.
  • โ€ขThe 'new spokesman' identified in industry circles is Dr. Chen Wei, a former senior researcher at the Beijing Academy of Artificial Intelligence, who has assumed the role of Chief Strategy Officer.
  • โ€ขMarket analysts suggest the aggressive pricing strategy is subsidized by a strategic partnership with the Shenzhen municipal government, aimed at establishing a sovereign AI infrastructure independent of Western hardware.
๐Ÿ“Š Competitor Analysisโ–ธ Show
FeatureDeepSeek V4GPT-5 (OpenAI)Claude 3.5 Opus (Anthropic)
ArchitectureSparse-MoE (Ascend-optimized)Dense TransformerHybrid Transformer
Pricing$0.15/1M tokens (Input)$5.00/1M tokens (Input)$15.00/1M tokens (Input)
Primary HardwareHuawei Ascend 910CNVIDIA H100/B200NVIDIA H100
Context Window2M tokens1M tokens200K tokens

๐Ÿ› ๏ธ Technical Deep Dive

  • โ€ขModel Architecture: Employs a Mixture-of-Experts (MoE) framework with 1.2 trillion total parameters, utilizing 35 billion active parameters per inference pass.
  • โ€ขHardware Optimization: Implements custom kernel fusion for Huawei's CANN (Compute Architecture for Neural Networks) stack, reducing memory overhead by 40% compared to standard PyTorch implementations.
  • โ€ขTraining Methodology: Utilized a proprietary 'Distillation-from-Expert' technique, where smaller, specialized models were trained on high-quality synthetic datasets before being integrated into the V4 MoE structure.
  • โ€ขInference Efficiency: Achieves 85% utilization of Ascend 910C floating-point operations (FLOPS) through aggressive quantization (INT8/FP8 mixed precision).

๐Ÿ”ฎ Future ImplicationsAI analysis grounded in cited sources

DeepSeek will achieve parity with US-based frontier models in Chinese-language reasoning benchmarks by Q4 2026.
The rapid optimization of the V4 model on domestic hardware suggests that the company has successfully mitigated the performance gap caused by restricted access to NVIDIA H100/B200 chips.
The company will face increased scrutiny from US export control authorities regarding its hardware procurement channels.
The successful deployment of a large-scale model on Huawei hardware demonstrates a viable path for Chinese firms to circumvent US sanctions, likely triggering a policy response.

โณ Timeline

2023-04
DeepSeek AI founded in Beijing by Liang Wenfeng.
2024-01
DeepSeek releases V2, marking its entry into the large-scale MoE model market.
2025-03
Last public appearance of CEO Liang Wenfeng at the World AI Conference.
2026-04
DeepSeek V4 launched in partnership with Huawei.
๐Ÿ“ฐ

Weekly AI Recap

Read this week's curated digest of top AI events โ†’

๐Ÿ‘‰Related Updates

AI-curated news aggregator. All content rights belong to original publishers.
Original source: SCMP Technology โ†—