AWS and Nvidia build a supercomputer with 16,384 Superchips, Team Up for Generative AI Infrastructure

Nvidia
(Image credit: Nvidia)

Although many companies are developing accelerators for artificial intelligence (AI) workloads, Nvidia's CUDA platform is currently unrivaled regarding AI support. As a result, demand for Nvidia-based AI infrastructure is high. To address it, Amazon Web Services and Nvidia entered a strategic partnership under which AWS will offer Nvidia-based infrastructure for generative AI. The two companies will partner on several key projects.

"Today, we offer the widest range of Nvidia GPU solutions for workloads including graphics, gaming, high performance computing, machine learning, and now, generative AI," said Adam Selipsky, CEO at AWS. "We continue to innovate with Nvidia to make AWS the best place to run GPUs, combining next-gen Nvidia Grace Hopper Superchips with AWS's EFA powerful networking, EC2 UltraClusters' hyper-scale clustering, and Nitro's advanced virtualization capabilities." 

In addition, AWS will be the first to offer a cloud-based AI supercomputer based on Nvidia's GH200 Grace Hopper Superchips. This unique configuration will connect 32 Grace Hopper Superchips per instance using NVLink. It will scale up to thousands of GH200 Superchips (and 4.5 TB HBM3e memory) connected with Amazon's EFA networking and supported by advanced virtualization (AWS Nitro System) and hyper-scale clustering (Amazon EC2 UltraClusters).

Anton Shilov
Contributing Writer

Anton Shilov is a contributing writer at Tom’s Hardware. Over the past couple of decades, he has covered everything from CPUs and GPUs to supercomputers and from modern process technologies and latest fab tools to high-tech industry trends.

  • bit_user
    *Yawn*

    So... how many Watts does it burn, at full load? I guess it might be interesting to know something about the pricing structure, as well.
    Reply
  • weber462
    Computational power has become new arms race. Keep it with the masses. Not just rich. Support distributive computing.
    Reply
  • bit_user
    weber462 said:
    Computational power has become new arms race. Keep it with the masses. Not just rich. Support distributive computing.
    I already told you - a lot of problems need tight integration of compute elements and don't map well to a distributed computing processing model.

    In a way, cloud computing is indeed very democratic. I could never access a supercomputer before, but now it's potentially within reach for me to rent a little time on one.
    Reply
  • thisisaname
    Generative AI the gift that keeps on taking, right now it seems to me a solution looking for a problem to solve.
    Reply
  • sygreenblum
    bit_user said:
    *Yawn*

    So... how many Watts does it burn, at full load? I guess it might be interesting to know something about the pricing structure, as well.
    That's a fair question. I just read a NY times and MIT article where AI is expected to exceed AC units as the largest single source requiring as much as 21 percent of the grid by 2030, which would likely be higher that projected EV consumption as well. Google AI/data centers already use more power than the country of Ireland.
    Reply