AMD Packs 432 GB Of HBM4 Into Instinct MI400 GPUs For Double-Wide AI Racks
Packing 432 GB of HBM4 memory, the Instinct MI400 series GPUs due next year will provide 50 percent more memory capacity and bandwidth than Nvidia’s upcoming Vera Rubin platform. They will power rack-scale solutions that offers the same GPU density as Vera Rubin’s rack.
AMD revealed that its Instinct MI400-based, double-wide AI rack systems coming next year will provide 50 percent more memory capacity and bandwidth than Nvidia’s upcoming Vera Rubin platform while offering roughly the same compute performance.
The chip designer shared the first details Thursday during AMD CEO Lisa Su’s keynote at its Advancing AI event in San Jose, Calif. The company also shared many more details for its MI350X GPUs and its corresponding products that will hit the market later this year.
[Related: 9 AMD Acquisitions Fueling Its AI Rivalry With Nvidia]
Andrew Dieckman, corporate vice president and general manager of data center GPU at AMD, said in a Wednesday briefing that the MI400 series “will be the highest-performing AI accelerator in 2026 for both large-scale training as well as inferencing.”
“It will bring leadership performance due to the memory capacity, memory bandwidth, scale-up bandwidth and scale up bandwidth advantages,” he said of the MI400 series and the double-wide “Helios” AI server rack it will power.
Set to launch in 2026, AMD said the MI400 GPU series will be capable of performing 40 petaflops of 4-bit floating point (FP4) and 20 petaflops of 8-bit floating point (FP8), double that of the flagship MI355X landing this year.
Compared to the MI350 series, the MI400 series will increase memory capacity to 432 GB based on the HBM4 standard, which will give it a memory bandwidth of 19.6 TBps, more than double that of the previous generation. The MI400 series will also sport a scale-out bandwidth capacity of 300 GBps per GPU.
AMD plans to pair the MI400 series with its next-generation EPYC “Venice” CPU and Pensando “Vulcano” NIC to power the Helios AI rack.
The Helios rack will consist of 72 MI400 GPUs, giving it 31 TB of HBM4 memory capacity, 1.4 PBps of memory bandwidth and 260 TBps of scale-up bandwidth. This will make it capable of 2.9 exaflops of FP4 and 1.4 exaflops of FP8. The rack will also have a scale-out bandwidth of 43 TBps.
Compared to Nvidia’s Vera Rubin platform that is set to launch next year, AMD said the Helios rack will come with the same number of GPUs and scale-up bandwidth plus roughly the same FP4 and FP8 performance.
At the same time, the company said Helios will offer 50 percent greater HBM4 memory capacity, memory bandwidth and scale-out bandwidth.
In a rendering of the Helios, the rack appeared to be wider than Nvidia’s rack-scale solutions such as the GB200 NVL72 platform.
Dieckmann said Helio is a double-wide rack, which AMD and its key partners felt was the “right design point between complexity [and] reliability.”
The executive said the rack’s wider size was a “trade-off” to achieve Helios’ performance advantages, including the 50 percent greater memory bandwidth.
“That trade-off was deemed to be a very acceptable trade-off because these large data centers tend not to be square footage constrained. They tend to be megawatts constrained. And so we think this is the right design point for the market based upon what we’re delivering,” Dieckmann said.