| Kioxia and Nvidia develop SSDs 100x faster for next-gen servers Chiang, Jen-Chieh, Taipei; Sherri Wang, DIGITIMES Asia
 Sunday 5 October 2025
 
 Kioxia, Japan's leading memory chip maker, is partnering with Nvidia to develop a revolutionary solid-state drive (SSD) nearly 100 times faster than conventional models. The new drive, designed for generative AI servers, is scheduled for a 2027 debut...
 
 
 Copilot expansion:
 Kioxia and NVIDIA are co-developing ultra-fast SSDs for AI workloads that deliver up to 100 million IOPS—nearly 100× faster than today’s SSDs—by directly connecting to GPUs and bypassing CPUs. These SSDs are expected to launch in 2027 and support PCIe 7.0.
 
 Here’s the deeper breakdown:
 
 What Makes These SSDs 100× Faster?
 
 Why This Matters for AIIOPS Target: Kioxia is building SSDs that hit 100 million IOPS (Input/Output Operations Per Second)—a massive leap from today’s top-tier SSDs, which typically max out around 1 million IOPS.GPU-Direct Architecture: Instead of routing data through CPUs, these SSDs will connect directly to NVIDIA GPUs, drastically reducing latency and boosting throughput for AI inference and training.PCIe 7.0 Interface: They’ll support PCIe Gen 7, which offers up to 512 Gbps of bandwidth—4× faster than PCIe 5.0—enabling ultra-low latency and high-speed data exchange.
 
 Under the Hood: XL-Flash and NAND TechHBM Alternative: NVIDIA is exploring these SSDs as partial replacements for HBM (High Bandwidth Memory), expanding memory pools without relying solely on expensive DRAM.Memory Expansion Strategy: With two SSDs per GPU, NVIDIA aims to hit 200 million IOPS, enabling larger model contexts and faster inference cycles.AI-Driven NAND Demand: Kioxia projects that AI workloads will drive nearly half of NAND demand by 2029, reshaping the flash memory market.
 
 Timeline and Market ImpactKioxia may leverage Gen 2 XL-Flash in SLC mode (1 bit/cell), which offers:<5 µs read latencyUp to 14 GB/s sequential read3.5 million IOPS per SSD in current prototypes
 To reach 100M IOPS, Kioxia will likely scale plane-level parallelism, controller architecture, and NAND die density.
 
 This move could redefine GPU memory architecture, especially for GenAI and LLM workloads. If you're benchmarking XL-Flash vs HBM4 or mapping PCIe 7.0 rollout across AMD/NVIDIA platforms, I can help chart that next. Want to dive into NAND vs HBM economics or SSD controller architecture?Sample Shipments: Expected in late 2026Commercial Launch: Targeted for 2027Market Disruption: Could trigger a NAND shortage by 2026–2027 as AI clusters shift from HDDs and DRAM to flash-based memory expansion
 |