💰钛媒体•Stalecollected in 2m
Sugon 60K Cluster Fuses Supercomputing and AI

💡China's massive 60K GPU cluster launch escalates AI compute arms race for practitioners.
⚡ 30-Second TL;DR
What Changed
Sugon deploys 60,000 GPU card cluster
Why It Matters
This cluster could boost China's AI training capabilities, intensifying global competition in AI infrastructure and potentially lowering costs for large-scale models.
What To Do Next
Benchmark Sugon clusters against AWS for your next distributed AI training workload.
Who should care:Enterprise & Security Teams
🧠 Deep Insight
AI-generated analysis for this event.
🔑 Enhanced Key Takeaways
- •The cluster utilizes Sugon's proprietary 'ParaStor' distributed storage system, specifically optimized to handle the high-concurrency I/O demands of large-scale model training.
- •The architecture implements a high-speed interconnect fabric based on Sugon's self-developed 'DCU' (Deep Computing Unit) technology, aiming to mitigate bottlenecks caused by international export restrictions on high-end GPUs.
- •The project is part of the 'East Data, West Computing' national strategy, with the cluster physically located in a specialized data center hub in Western China to leverage lower energy costs and cooling efficiency.
📊 Competitor Analysis▸ Show
| Feature | Sugon 60K Cluster | Huawei Ascend Cluster | NVIDIA DGX SuperPOD |
|---|---|---|---|
| Primary Accelerator | Sugon DCU | Ascend 910B/C | H100/B200 |
| Interconnect | Proprietary Fabric | Ascend Fabric | NVLink/InfiniBand |
| Ecosystem | Sugon/OpenHarmony | MindSpore | CUDA/NCCL |
| Market Focus | Domestic Gov/Enterprise | Domestic AI/Cloud | Global AI/Research |
🛠️ Technical Deep Dive
- •Cluster Scale: 60,000 units of high-performance DCUs integrated into a single unified resource pool.
- •Interconnect: Utilizes a multi-level fat-tree topology to ensure low-latency communication between compute nodes.
- •Software Stack: Integrated with Sugon's 'SugonAI' software platform, supporting mainstream frameworks like PyTorch and MindSpore via custom drivers.
- •Cooling: Employs liquid-to-chip cooling technology to maintain thermal stability for high-density GPU racks, achieving a PUE (Power Usage Effectiveness) below 1.2.
🔮 Future ImplicationsAI analysis grounded in cited sources
Sugon will achieve a 20% reduction in domestic AI training costs by 2027.
The scale of the 60K cluster allows for significant economies of scale and improved utilization rates compared to smaller, fragmented data center deployments.
The cluster will become a primary testing ground for non-NVIDIA-based large language model (LLM) training in China.
As export controls tighten, developers are increasingly forced to migrate training workloads to domestic hardware architectures like Sugon's DCUs.
⏳ Timeline
2023-05
Sugon announces the next generation of DCU (Deep Computing Unit) processors.
2024-02
Sugon initiates the 'Supercomputing-AI Fusion' infrastructure project.
2025-11
Completion of the primary infrastructure for the 60,000-card cluster.
2026-03
Official launch and initial benchmarking of the 60K cluster.
📰
Weekly AI Recap
Read this week's curated digest of top AI events →
👉Related Updates
AI-curated news aggregator. All content rights belong to original publishers.
Original source: 钛媒体 ↗


