Nvidia launches Vera Rubin NVL72 AI supercomputer at CES — promises up to 5x greater inference performance and 10x lower cost per token than Blackwell, coming 2H 2026

Nvidia Vera Rubin, CES 2026
(Image credit: Tom's Hardware)

AI is everywhere at CES 2026, and Nvidia GPUs are at the center of the expanding AI universe. Today, during his CES keynote, CEO Jensen Huang shared his plans for how the company will remain at the forefront of the AI revolution as the technology reaches far beyond chatbots into robotics, autonomous vehicles, and the broader physical world.

First up, Huang officially launched Vera Rubin, Nvidia's next-gen AI data center rack-scale architecture. Rubin is the result of what the company calls "extreme co-design" across six types of chips: the Vera CPU, the Rubin GPU, the NVLink 6 switch, the ConnectX-9 SuperNIC, the BlueField-4 data processing unit, and the Spectrum-6 Ethernet switch. Those building blocks all come together to create the Vera Rubin NVL72 rack.

Google Preferred Source

Follow Tom's Hardware on Google News, or add us as a preferred source, to get our latest news, analysis, & reviews in your feeds.

Jeffrey Kampman
Senior Analyst, Graphics

As the Senior Analyst, Graphics at Tom's Hardware, Jeff Kampman covers everything to do with GPUs, gaming performance, and more. From integrated graphics processors to discrete graphics cards to the hyperscale installations powering our AI future, if it's got a GPU in it, Jeff is on it. 

  • Stomx
    What about fp32 and fp64? Will these be included in every chip or there will be special versions of this GPU for HPC and supercomputers?
    Reply
  • timsSOFTWARE
    This is also the problem with these companies stockpiling hardware. It's a depreciating asset, and becomes effectively worthless after 5 years or so, because of the energy cost and space required vs. replacement hardware. Nvidia is saying this new gen will be 10x more energy efficient than the one before it. If that happens again in a few more years - what will Blackwell hardware be worth, when the new stuff can fit in 5% of the space and use 1% of the power?

    And the majority of the AI industry seems to be betting on having a use for continued scale - and relatively quick results - but I agree with various industry voices that are saying scale is pretty much played out already, and additional improvements are probably going to take some time - the limitation now is more about ideas than hardware.
    Reply
  • bit_user
    timsSOFTWARE said:
    Nvidia is saying this new gen will be 10x more energy efficient than the one before it.
    You have to take their numbers with a grain of salt. They're pros at cherry-picking and using every available trick to make their numbers as big as possible.

    timsSOFTWARE said:
    If that happens again in a few more years - what will Blackwell hardware be worth, when the new stuff can fit in 5% of the space and use 1% of the power?
    They were focused on inferencing. The improvements for training will be much less. Hence, some existing hardware that's used for inferencing could be re-targeted towards training and still remain viable for perhaps a couple more generations.

    That said, I agree that this hardware has a limited window for seeing a return-on-investment and there's barely enough power for datacenters as-is. It seems like they'll be forced to decommission some not-so-old hardware, in order to make room & power/cooling budgets for new stuff.

    Some macroeconomic butterfly is going to flap its wings, somewhere, and the cascading effect through the economy will squeeze these guys by shutting off their funding to do more GPU buys and datacenter buildouts. Then, things are going to get interesting - especially if some of the biggest AI companies are unable even to service their debts.
    Reply
  • alan.campbell99
    Demand for AI compute is insatiable, based on what, how much debt OpenAI and Anthropic are taking on? I've seen reporting that out of 400+ million paid 365 users only 8 million are paying for Copilot.
    Reply