🏠IT之家•Freshcollected in 8m
Dell: AI Memory Demand 625x by 2028

💡AI memory to 625x by '28—supply crisis ahead, secure Dell infra now
⚡ 30-Second TL;DR
What Changed
625x total memory growth: 25x capacity x 25x accelerators vs 2023
Why It Matters
Explosive demand signals infrastructure crunch, raising costs for AI training/inference. Enterprises must secure supply chains early; Dell positions as reliable vendor amid shortages.
What To Do Next
Reach out to Dell for AI server RFQs to lock in memory-heavy configs before 2028 shortages.
Who should care:Enterprise & Security Teams
🧠 Deep Insight
AI-generated analysis for this event.
🔑 Enhanced Key Takeaways
- •The 2TB per-accelerator memory target is heavily reliant on the transition from HBM3e to HBM4 and HBM4e, which integrate logic dies directly onto the memory stack to improve bandwidth-per-watt efficiency.
- •Dell's strategy involves shifting from standard air-cooled server architectures to direct-to-chip liquid cooling, which is a prerequisite for managing the thermal density of the high-capacity memory stacks required by 2028.
- •Industry analysts note that the 625x growth projection assumes a massive shift toward 'inference-at-the-edge' and 'agentic AI' models, which require significantly larger local context windows than the training-focused models of 2023.
📊 Competitor Analysis▸ Show
| Feature | Dell (AI Server Strategy) | HPE (ProLiant Gen11/12) | Supermicro (AI Solutions) |
|---|---|---|---|
| Cooling Focus | Direct-to-chip liquid cooling | Liquid cooling / Immersion | High-density air / Liquid |
| Supply Chain | Tier-1 vendor partnerships | Strategic silicon alliances | Agile, modular component sourcing |
| Market Positioning | Enterprise/Data Center scale | HPC/Supercomputing focus | Rapid deployment/Custom builds |
🛠️ Technical Deep Dive
- •Memory Scaling: Transitioning from 8-high HBM3e stacks to 12-high and 16-high HBM4 stacks to achieve the 2TB per-accelerator capacity.
- •Interconnect Architecture: Implementation of NVLink Switch systems and PCIe Gen6/7 to handle the increased data movement required by the expanded memory footprint.
- •Thermal Management: Adoption of coolant distribution units (CDUs) and cold plates capable of dissipating 100kW+ per rack to support the high-TDP accelerators.
- •Memory Hierarchy: Integration of CXL (Compute Express Link) 3.0/3.1 to allow for memory pooling and expansion beyond the physical limits of the accelerator's local HBM.
🔮 Future ImplicationsAI analysis grounded in cited sources
HBM supply will remain the primary bottleneck for AI server revenue through 2027.
The 4-year lead time for new DRAM fab capacity prevents supply from scaling linearly with the exponential demand for high-capacity memory stacks.
Dell will achieve a higher average selling price (ASP) per server by 2028.
The shift to high-capacity HBM and complex liquid cooling infrastructure significantly increases the bill-of-materials cost per unit.
⏳ Timeline
2023-05
Dell announces Project Helix in partnership with NVIDIA to simplify generative AI deployment.
2024-03
Dell expands AI server portfolio with PowerEdge XE9680, optimized for high-performance AI workloads.
2025-02
Dell reports record AI-optimized server demand, signaling a shift in revenue composition toward AI infrastructure.
📰
Weekly AI Recap
Read this week's curated digest of top AI events →
👉Related Updates
AI-curated news aggregator. All content rights belong to original publishers.
Original source: IT之家 ↗
