Nvidia confirms Blackwell Ultra and Vera Rubin GPUs are on track for 2025 and 2026 — post-Rubin GPUs in the works

Nvidia
(Image credit: Nvidia)

A design flaw delayed Nvidia's rollout of Blackwell GPUs for data centers last year, prompting the company to redesign the silicon and packaging. However, this did not affect Nvidia's work on mid-cycle refresh Blackwell 300-series (Blackwell Ultra) GPUs for AI and HPC and next-generation Vera Rubin GPUs. Nvidia can already share some details about Rubin GPUs and its post-Rubin products.

"Blackwell Ultra is [due in the] second half," Jensen Huang, chief executive of Nvidia, reaffirmed analysts and investors at the company's earnings conference call. "The next train [is] Blackwell Ultra with new networking, new [12-Hi HBM3E] memory, and, of course, new processors. […] We have already revealed and have been working very closely with all of our partners on the click after that. The click after that is called Vera Rubin and all of our partners are getting up to speed on the transition to that. […] [With Rubin GPUs] we are going to provide a big, big, huge step up."

Later this year, Nvidia plans to release its Blackwell B300-series solutions for AI and HPC (previously known as Blackwell Ultra) that will offer higher compute performance as well as eight stacks of 12-Hi HBM4E memory, thus providing up to 288GB of memory onboard. Unofficial information indicates that performance uplift enabled by Nvidia's B300-series will be around 50% compared to the comparable B200-series products, though the company has yet to confirm this.

To further improve performance, B300 will be offered with Nvidia's Mellanox Spectrum Ultra X800 Ethernet switch, which has a radix of 512 and can support up to 512 ports. Nvidia is also expected to provide additional system design freedom to its partner with its B300-series data center GPUs.

(Image credit: Constellation Research)

Nvidia's next-generation GPU series will be based on the company's all-new, codenamed Rubin architecture, further improving AI compute capabilities as industry leaders march towards achieving artificial general intelligence (AGI). In 2026, the first iteration of Rubin GPUs for data centers will come with eight stacks of HBM4E memory (up to 288GB). The Rubin platform will also include a Vera CPU, NVLink 6 switches at 3600 GB/s, CX9 network cards supporting 1,600 Gb/s, and X1600 switches.

Jensen Huang plans to talk about Rubin at the upcoming GPU Technology Conference (GTC) in March, though it remains to be seen what he plans to discuss. Surprisingly, Nvidia also intends to talk about post-Rubin products at the GTC. From what Jensen Huang announced this week, it is unclear whether the company plans to reveal details of Rubin Ultra GPUs or its GPU architecture that will come after the Rubin family.

Speaking of Rubin Ultra, this could indeed be quite a breakthrough product. It is projected to come with 12 stacks of HBM4E in 2027 once Nvidia learns how to efficiently use 5.5-reticle-size CoWoS interposers and 100mm × 100mm substrates made by TSMC.

"Come to GTC and I will [tell you about] Blackwell Ultra," said Huang. "There are Rubin, and then [we will] show you what is one click after that. Really, really exciting new product."

Anton Shilov
Contributing Writer

Anton Shilov is a contributing writer at Tom’s Hardware. Over the past couple of decades, he has covered everything from CPUs and GPUs to supercomputers and from modern process technologies and latest fab tools to high-tech industry trends.

Read more
Nvidia data center GPU roadmap 2025 showing Rubin and Rubin Ultra
Nvidia announces Rubin GPUs in 2026, Rubin Ultra in 2027, Feynman also added to roadmap
Nvidia Rubin Ultra with NVL576 Kyber racks and infrastructure
Nvidia shows off Rubin Ultra with 600,000-Watt Kyber racks and infrastructure, coming in 2027
Nvidia Blackwell Ultra B300
Nvidia announces Blackwell Ultra B300 —1.5X faster than B200 with 288GB HBM3e and 15 PFLOPS dense FP4
Blackwell
Nvidia's next-gen B300 GPUs have 1,400W TDP, deliver 50% more AI horsepower: Report
Dell servers based on Nvidia GB200
Nvidia says Blackwell-based servers are in full production - 200 different configurations now available
Nvidia Blackwell Architecture deep dive slides
Nvidia Blackwell architecture deep dive: A closer look at the upgrades coming with RTX 50-series GPUs
Latest in GPUs
Nvidia Ada Lovelace and GeForce RTX 40-Series
Nvidia is reportedly close to adopting Intel Foundry's 18A process node for gaming GPUs
RX 9070 XT Sapphire
Lisa Su says Radeon RX 9070-series GPU sales are 10X higher than its predecessors — for the first week of availability
RTX 5070, RX 9070 XT, Arc B580
Real-world GPU prices cost up to twice the MSRP — a look at current FPS per dollar values
Zotac Gaming GeForce RTX 5090 AMP Extreme Infinity
Zotac raises RTX 5090 prices by 20% and seemingly eliminates MSRP models
project-g-assist-nvidia-geforce-rtx-ogimage
Nvidia releases public G-Assist in latest App to provide in-game AI assistance — also introduces DLSS custom scaling factors
Yeston Sakura Radeon RX 9070 XT
This scent-dispensing RX 9070 XT assures at least one GPU launch this year doesn't stink
Latest in News
Nvidia Ada Lovelace and GeForce RTX 40-Series
Nvidia is reportedly close to adopting Intel Foundry's 18A process node for gaming GPUs
RX 9070 XT Sapphire
Lisa Su says Radeon RX 9070-series GPU sales are 10X higher than its predecessors — for the first week of availability
RTX 5070, RX 9070 XT, Arc B580
Real-world GPU prices cost up to twice the MSRP — a look at current FPS per dollar values
Zotac Gaming GeForce RTX 5090 AMP Extreme Infinity
Zotac raises RTX 5090 prices by 20% and seemingly eliminates MSRP models
ASRock fixes AM5 motherboard by cleaning it
ASRock claims to fix 'burned out' AM5 motherboard by cleaning the socket
ChatGPT Security
Some ChatGPT users are addicted and will suffer withdrawal symptoms if cut off, say researchers
  • bit_user
    I wonder when they're going to begin moving away from CUDA. I think it really is holding them back, when most others have already move on to dataflow architectures. I liken Nvidia's current lead to how Intel's lead from around 10-15 years ago was primarily based on better manufacturing tech.

    CUDA was once Nvidia's biggest advantage, but now it's turning into their greatest liability. They know it, too. Just look at their edge SoCs, which pack most of their inferencing horsepower in the NVDLA engines, not the iGPUs.
    Reply
  • George³
    HBM4E, hmm, why Nvidia have not plans to use 16 stacks?
    Reply
  • bit_user
    George³ said:
    HBM4E, hmm, why Nvidia have not plans to use 16 stacks?
    Just a few guesses:
    Perhaps the timeframe for availability of the 12-high capable dies is sooner?
    Heat, given that it's going to be stacked atop logic dies. Heat was also an issue for them in B200: https://www.tomshardware.com/tech-industry/artificial-intelligence/nvidia-partners-indirectly-confirms-blackwell-b200-gpu-delay-offer-interested-parties-liquid-cooled-h200-instead Maybe higher stacks require more vias, making it less cost- & area-efficient?
    Perhaps the optimal compute vs. data ratio favors slightly smaller stacks?
    Or, perhaps something to do with concerns about yield, the more dies you stack. Yield seems to have been an ongoing area of struggle, for the B200.
    Reply
  • TCA_ChinChin
    bit_user said:
    I wonder when they're going to begin moving away from CUDA. I think it really is holding them back, when most others have already move on to dataflow architectures. I liken Nvidia's current lead to how Intel's lead from around 10-15 years ago was primarily based on better manufacturing tech.

    CUDA was once Nvidia's biggest advantage, but now it's turning into their greatest liability. They know it, too. Just look at their edge SoCs, which pack most of their inferencing horsepower in the NVDLA engines, not the iGPUs.
    What do you mean? The current industry seems entrenched in CUDA and even if there are cases where different architectures might have advantages, I can't see CUDA moving away or even hindering Nvidia's position much in the future. The only competitors I'm aware of like ROCm and UXL aren't really very competitive.

    Nvidia can certainly fumble like Intel has done recently but I wouldn't say CUDA is a liabilty now, just like 15 years ago I never though Intel would be in their current position.
    Reply
  • valthuer
    AI/Datacenters, is where real money is made for Nvidia. I wonder when they 're gonna stop making consumer GPUs.
    Reply
  • bit_user
    TCA_ChinChin said:
    What do you mean?
    CUDA is too general and depends on SMT to achieve good utilization. SMT isn't the most efficient, in terms of energy or silicon area. Dedicated NPUs don't work this way - they use local memory, DSP cores, and rely on DMAs to stream data in and out of local memory. Nvidia has been successful in spite of CUDA's inefficiency, but they can't keep their lead forever if they don't switch over to a better programming model for the problem.

    They did succeed in tricking AMD into following them with HIP, rather than potentially leap-frogging them. With oneAPI, I think Intel is also basically following the approach Nvidia took with CUDA. As long as they keep thinking they're going to beat Nvidia at its own game, they deserve to keep losing. AMD should've bought Tenstorrent or Cerebras, but now they're probably too expensive.

    At least AMD and Intel both have sensible inferencing hardware. Lisa Su should pay attention to the team at Xilinx that designed what they now call XDNA - I hope that's what UDNA is going to be.

    TCA_ChinChin said:
    The current industry seems entrenched in CUDA and even if there are cases where different architectures might have advantages, I can't see CUDA moving away or even hindering Nvidia's position much in the future.
    I already gave you an example where even Nvidia clearly sees the light. Just look at their NVDLA engines, which are now already on their second generation. Those aren't programmable using CUDA.

    Just sit back and wait. I wonder if Rubin is going to be their first post-CUDA training architecture.
    Reply
  • hannibal
    valthuer said:
    AI/Datacenters, is where real money is made for Nvidia. I wonder when they 're gonna stop making consumer GPUs.

    They just start making gaming gpus by using older nodes that are not needed for AI chips!
    $2000 to $ 4000 for gaming GPU is still profit even if it is tiny compared to AI stuff. So don´t worry! There will be overpriced gaming GPUs also in the future!

    ;)
    Reply
  • valthuer
    hannibal said:
    There will be overpriced gaming GPUs also in the future!

    ;)

    Thank God! :ROFLMAO:
    Reply
  • bit_user
    hannibal said:
    They just start making gaming gpus by using older nodes that are not needed for AI chips!
    RTX 5090 is made on a TSMC N4-class node, which is the same as they're using for AI training GPUs. The RTX 5090's die can't get much bigger. So, the only way they could do a 3rd generation on this family of nodes is by going multi-die, which Intel and AMD have dabbled with (and Apple successfully executed), but Nvidia has steadfastly avoided. What I've heard about the multi-die approach is that the amount of global data movement makes this inefficient for rendering. So, I actually doubt they'll go in that direction, at least not yet.

    My prediction is that they'll use a 3 nm-class node for their next client GPU. They're already set to use a N3 node for Rubin, later this year.

    hannibal said:
    $2000 to $ 4000 for gaming GPU is still profit even if it is tiny compared to AI stuff. So don´t worry! There will be overpriced gaming GPUs also in the future!
    Yeah, I think the main risk that AI poses to their gaming products is simply that it tends to divert resources and focus. That's probably behind some of the many problems that have so far affected the RTX 5000 launch.

    Nvidia does seem to keep doing research on things like neural rendering, so that clearly shows they're not leaving graphics any time soon. It's more that they're focusing on the intersections between AI and graphics, which is certainly better than nothing.
    Reply
  • TCA_ChinChin
    bit_user said:
    I already gave you an example where even Nvidia clearly sees the light. Just look at their NVDLA engines, which are now already on their second generation. Those aren't programmable using CUDA.

    Just sit back and wait. I wonder if Rubin is going to be their first post-CUDA training architecture.
    I can see Nvidia introducing new architectures in the future and shifting away from CUDA eventually, I can't see Nvidia's position dominance being challenged. My point is:

    "Nvidia can certainly fumble like Intel has done recently but I wouldn't say CUDA is a liabilty now, just like 15 years ago I never though Intel would be in their current position."

    Despite all the drawbacks you've pointed out, they're clearly still on top of their game. CUDA has technical downsides but it certainly doesn't have significant financial downsides for Nvidia, at least not yet. I'm not holding my breath for Rubin, maybe afterwards.
    Reply