AMD preps rack-scale Instinct MI450X IF128 with 128 GPUs to challenge Nvidia's VR200 NVL144 in 2026
AMD's venture carries significant complexity and deployment risks.

AMD plans to launch its first two rack-scale Instinct accelerators in 2026 to compete with Nvidia's VR300 NVL144, reports SemiAnalysis. They are named AMD's Instinct MI450X IF64 and Instinct MI450X IF128, and both are designed for AI deployment. If they prove to be a success, it could change the landscape of AI hardware over time.
While AMD's Instinct MI300-series AI and HPC GPUs are very powerful on paper, they cannot compete against Nvidia's GB200 NVL72 rack-scale solution in terms of performance scalability, as their maximum scale-up world size is eight processors.
But next year, things are going to change as AMD plans to release its Instinct MI450X IF64, and Instinct MI450X IF128 solutions, replete with 64 and 128 GPU packages, to compete against Nvidia's VR200 NVL144 (with 72 GPU packages).
Theoretically, AMD's MI450X IF128 can have an edge over Nvidia's VR200 NVL144. However, its complexity and technical challenges may limit its initial success.
The Instinct MI450X IF128 will be AMD's first system to support multiple AI processors across two racks using Infinity Fabric, extended over Ethernet. The machine will rely on 16 1U servers running one AMD EPYC 'Venice' CPU with four Instinct MI450X GPUs equipped with their own LPDDR memory pool and a PCIe x4 SSD. Each of the 128 GPUs will have over 1.8 TB/s of unidirectional internal bandwidth for inter-GPU communication within the same scaling domain, thus enabling significantly larger compute clusters than AMD has supported so far.
For scale-out communication outside the local group of GPUs (i.e., MI450X IF128 machines), the system will include up to three 800GbE Pensando network cards for each GPU. This provides a total outbound network bandwidth of 2.4 Tb/s per device (via PCIe). A secondary configuration will also be available, allowing each GPU to use two 800GbE network cards connected using a PCIe interface. However, this version will not be able to use the full bandwidth of the interfaces, as the PCIe 5.0 links are insufficient to fully support two high-speed network cards.
Unlike Nvidia's GB200-series systems, which use active optical cables with embedded components to connect racks, AMD will employ a simpler passive copper wiring approach. This strategy may help reduce system cost and power consumption, but could be limited by signal integrity or cable length constraints.
Stay On the Cutting Edge: Get the Tom's Hardware Newsletter
Get Tom's Hardware's best news and in-depth reviews, straight to your inbox.
Also, due to the system's complexity, manufacturing and deployment may face delays or technical issues. To address this risk, AMD is preparing a smaller version of the same architecture called MI450X IF64. This variant will be confined to a single rack and use a simplified interconnect design, which promises to enable a more predictable rollout.
If AMD manages to execute this architecture successfully, it could improve its position in the AI compute market, particularly AI inference systems. Whether it will be able to challenge Nvidia is something that remains to be seen, though.
Follow Tom's Hardware on Google News to get our up-to-date news, analysis, and reviews in your feeds. Make sure to click the Follow button.

Anton Shilov is a contributing writer at Tom’s Hardware. Over the past couple of decades, he has covered everything from CPUs and GPUs to supercomputers and from modern process technologies and latest fab tools to high-tech industry trends.