Skip to main content

Nvidia GTC 2026 keynote live blog — Vera Rubin GPUs and CPUs, DLSS 5, and the 'future of technology'

A lot of AI announcements, and maybe a thing or two for consumers.

GTC 2026
(Image credit: © Tom's Hardware)

Nvidia's GTC 2026 keynote has wrapped. During the two-hour (and change) presentation, Nvidia CEO Jensen Huang delivered several announcements, from new Vera CPUs to Groq LPUs, as well as laid out a vision for AI over the next 12 months.

As with any Nvidia event these days, we heard a lot about AI, as well as learned about Vera Rubin systems and DLSS 5. See our full live blog below.

Refresh

We're moments away from GTC 2026

What to expect from GTC 2026

  • Intel x Nvidia partnership — Nvidia bought $5 billion in Intel stock last year, and at the time, announced that the two companies would be working together on custom x86 processors across both the data center and consumer PCs. The deal has apparently been decades in the making. It's not clear if we'll hear about consumer or enterprise chips, or both, but there's a good chance we'll hear something from the partnership.
  • The 'future of real-time rendering' Nvidia presented at GDC (not GTC) about neural rendering, but just a week later, the company is teasing that it will reveal the "future of real-time rendering" at GTC (not GDC). Maybe it's a new DLSS feature, maybe it's something completely new. We don't know, but Nvidia has already confirmed something is coming for gamers during the keynote.
  • More on Vera Rubin Nvidia officially launched its Vera Rubin NVL72 in January, and it started shipping samples to customers just weeks ago. These next-gen AI data center boards are on-track for the second half of the year, so we expect to hear a lot about them during the keynote.
  • AI agents — Since the release of OpenClaw, the tech industry has been washed in talk of AI agents. Nvidia will talk about AI agents during the keynote, that much is almost guaranteed. We could see an announcement of "NemoClaw," which is an AI agent Nvidia is reportedly developing to compete with OpenClaw.
  • Nvidia N1/N1X — Perhaps the biggest rumor around Nvidia over the past year has been the N1 and N1X, which are two SoCs reportedly being developed for the consumer market. Do we finally see a reveal at this year's GTC? Perhaps, but this is the last item on this list for a reason.

T-Minus 5 Minutes

The GTC 2026 keynote stage.

(Image credit: Tom's Hardware.)

Run to the bathroom, get your drink ready, and settle in. We're just a few minutes away from the start of GTC 2026. Jensen will probably start with a short history of Nvidia's role in AI, but we expect the announcements to rapid-fire out after that point. We're sat down in the SAP Center in San Jose and ready to dig in.

Running a bit behind schedule

NVIDIA GTC Keynote 2026 - YouTube NVIDIA GTC Keynote 2026 - YouTube
Watch On

We're a few minutes past the top of the hour, and we're still waiting on the keynote to start. In the meantime, a quick reminder that you can watch along with us through the live stream above.

That's... a lot of country music?

And we're off!

The man of the hour is here: CEO Jensen Huang

Nvidia CEO Jensen Huang on stage at GTC 2026.

(Image credit: Tom's Hardware)

Jensen Huang has taken the stage in a familiar leather jacket. Sorry folks, there's no special jacket this time around. Jensen is starting off the show thanking some of the people that hosted the preshow leading up to the keynote.

'We've been working on CUDA for 20 years'

Nvidia CEO at GTC 2026.

(Image credit: Tom's Hardware)

CUDA is one of the major reasons Nvidia is in the position it's in today, and this GTC marks the 20th anniversary of CUDA. "The single hardest thing is to have built up our install base, we're in every cloud and computer company in every single industry," says Jensen.

Pricing of Ampere in the cloud is going up

'GeForce is Nvidia's greatest marketing campaign'

Nvidia CEO at GTC 2026.

(Image credit: Tom's Hardware)

Jensen says that "GeForce is Nvidia's greatest marketing campaign." It's an interesting way to frame the conversation, and one that Nvidia has been trying to crack for the past few years. Jensen paints a picture of Nvidia creating the first programmable shader 25 years ago, which eventually led to CUDA, and used GeForce as a vehicle to drive adoption.

Nvidia is showing off the next generation of computer graphics: DLSS 5

What is DLSS 5?

Nvidia presenting DLSS 5.

(Image credit: Tom's Hardware)

Nvidia says it combined controllable 3D graphics and structured data with generative worlds. "This concept of fusing structured data with generative AI will repeat itself in one industry after another industry after another industry."

'This is my best slide'

Nvidia CEO presenting a slide on structured data.

(Image credit: Future)

Jensen jokes that he's going to spend the rest of the keynote going through the slide you can see above about structured data. This is "the ground truth" of enterprise computing.

AI can solve unstructured data, says Jensen

More Moore's Law talk

Nvidia CEO talking about Google Cloud.

(Image credit: Tom's Hardware)

Jensen likes to talk about the death of Moore's Law, and he's doing so once again. "Moore's Law has run out of steam, accelerated computing allows us to take giant leaps forward." Jensen is showing off an example with Google Cloud and showing how Nvidia's acceleration can be repeated across companies and industries.

Nvidia is bringing OpenAI to AWS this year

GTC 2026

(Image credit: Tom's Hardware)

"As you know, [OpenAI] is completely compute-constrained." Jensen says that OpenAI will come to AWS this year, hopefully lightening the load on its massive infrastructure demand.

'Vertically integrated but horizontally open'

GTC 2026

(Image credit: Tom's Hardware)

Jensen describes Nvidia as "vertically integrated but horizontally open," which may or may not raise some eyebrows at the FTC. Regardless, Nvidia says there's "no other way" it can be given what it's trying to do with accelerated computing, delivering the entire stack to customers.

Nvidia says it needs domain-specific libraries to address the needs of different industries

GTC 2026

(Image credit: Tom's Hardware)

AI has a lot of applications, but Jensen says it isn't as simple as throwing GenAI at the wall and hoping it sticks. "We have to have domain-specific libraries that solve problems in every one of these verticals," he says.

Bringing it back to CUDA

cuDNN is what caused the 'big bang' of AI

A video demo at GTC 2026.

(Image credit: Tom's Hardware)

Nvidia says cuDNN, or CUDA Deep Neural Network, is one of the most important libraries the company has ever made, saying it caused the "big bang" of modern AI. Nvidia is showing a short video about its various CUDA-X libraries, including a life-like video that's entirely simulated.

Nvidia 'reinvented computing'

An accelerated timeline of AI

GTC 2026

(Image credit: Tom's Hardware)

There's been rapid AI development over the past few years. In 2023, it was ChatGPT. In 2024, it was reasoning models like o1, and in 2025, it was huge models with massive context windows like Claude Code. It's the first "agentic model," says Jensen. The executive says 100% of Nvidia is using Claude Code, along with other models. In 2026, Nvidia says we've reached an "inflection point for inference."

Nvidia says it's going to double demand through the next year

GTC 2026

(Image credit: Tom's Hardware)

Last year, Nvidia said it saw about $500 billion of high confidence demand and purchase orders for Blackwell and Rubin through 2026. "I see through 2027 at least $1 trillion," says Jensen. "Now, does it make any sense?" Jensen says that's what he's going to spend the rest of the keynote talking about.

Nvidia is the only company that runs every domain of AI across every domain of AI models

Nvidia says Grace Blackwell was 'a giant bet'

Nvidia CEO presenting at GTC 2026.

(Image credit: Tom's Hardware)

Nvidia NVL72 was a "giant bet," says Jensen, and he thanked Nvidia's partners for sticking with the company. "It wasn't easy for anybody... inference is the ultimate hard." The bet paid off, according to Nvidia, which you can see in the slide above.

50x performance per watt, 35x lower cost

'It's now a factory to generate tokens'

Nvidia CEO Jensen Huang presenting at GTC 2026.

(Image credit: Tom's Hardware)

Jensen says data centers used to be a place to store files, and they're now a factory to generate tokens. Inference is the workload and tokens are the new commodity, says Nvidia. Now, onto a short video showing how we got here.

Vera Rubin joins Jensen on stage

Vera Rubin at GTC 2026.

(Image credit: Tom's Hardware)

Vera Rubin NVL72 is the "engine supercharging the era of agentic AI." A new addition is the Groq 3 LPX tray, and as a whole, Nvidia says it's delivered 40 million times more compute over the past decade. Jensen is showing off Vera Rubin on stage; the whole thing. It's "one giant system."

Jensen says the Vera CPU is designed for high single-threaded performance, and the company built it go along with its racks for agentic processing.

Learn more about Nvidia's Vera CPU

Groq 3 LPU and Groq LPX join the fray

Nvidia CEO showing off Vera Rubin at GTC 2026.

(Image credit: Tom's Hardware)

A new addition to the system is a Groq LPX rack, which we learned about ahead of GTC. You can read Jeffrey Kampman's breakdown of Groq 3 in Vera Rubin now.

Nvidia GTC 2026

(Image credit: Tom's Hardware)

Jensen explains how NVLink for Rubin Ultra works, with compute sitting in the front and the scale-up fabric in the back.

This is 'the most important chart' for companies, says Nvidia

Nvidia CEO showing a chart at GTC 2026.

(Image credit: Tom's Hardware)

Tokens are "the new commodity," according to Nvidia. For businesses, Nvidia says that the throughput of an AI factory at iso power is something that will be "studied for years." More tokens means smarter models, and the smarter the models get, you need better token throughput. Nvidia says that, at every tier, Vera Rubin delivers much higher throughput.

Low latency and high throughput are 'enemies of each other'

Nvidia presenting the Groq 3 LPU at GTC 2026.

(Image credit: Nvidia)

Groq is important for Nvidia because it pushes beyond the limits of NVL72. With Groq LPX, Nvidia says it's able to deliver up to 10x in revenue to companies using Vera Rubin. It helps solve the problem of delivering low latency and high throughput, which Jensen described as "enemies of each other."

Nvidia combined one chip for high throughput and one for low latency, which it achieved with disaggregated inference.

Vera Rubin sampling is going 'incredibly well'

Vera Rubin is 7 chips across 5 rack systems

GTC 2026

(Image credit: Tom's Hardware)

Vera Rubin is undoubtedly Nvidia's most ambitious system to date, featuring seven chips across five rack systems. Compared to x86 and Hopper, Nvidia says Vera Rubin is able to deliver 700 million tokens per second compared to just 2 million

Here's Nvidia roadmap

GTC 2026

(Image credit: Tom's Hardware)

Jensen is teasing next-gen Feynman systems. It has a new GPU, new LPU, new CPU called Rosa, Bluefield 5, and Kyber with copper and CPO scale up. Feynman systems are on-track for 2028, so we'll hear a lot more about them throughout the year. At next year's GTC, we'll probably run back the same talking point with Feynman that we hard about with Vera Rubin this year.

Meet me in the Omniverse

Nvidia CEO presenting at GTC 2026.

(Image credit: Nvidia)

Nvidia built Omniverse to meet suppliers virtually, allowing co-design in the data center at a much broader scale. The goal is to leave "no power squandered." They're blueprints for AI factories, which Nvidia calls its DSX platform.

Data centers are going into space

Vera Rubin Space-1 at GTC 2026.

(Image credit: Tom's Hardware)

Nvidia is working on a system called Vera Rubin Space-1, which will be the first data center in space. Sounds like we're in early stages, but Nvidia has "a lot of great engineers" working on it.

NemoClaw makes using OpenClaw easy

Nemoclaw at Nvidia GTC 2026.

(Image credit: Tom's Hardware)

Nvidia is streamlining the process of setting up an AI agent with OpenClaw. Type two lines of shell commands, and you're off to the races with an AI agent. From there, Nvidia says you just need to give it a task and let the agent run its course.

What is OpenClaw? Nvidia says it's an OS

Nvidia worked with OpenClaw to make it enterprise-secure

Nvidia is building a Nemotron coalition

GTC 2026

(Image credit: Tom's Hardware)

Nvidia says that Nemotron 3 Ultra will be the best base model in the world. In order to scale out Nemotron, Nvidia is creating a coalition for Nemotron 4, including companies like Black Forest Labs, Perplexity, Mistral, and Cursor.

Bringing agents to the physical world

GTC 2026

(Image credit: Tom's Hardware)

Nvidia is showing off 110 robots at GTC, showcasing its "physical AI." Nvidia announced several new partners, including four new partners for robo-taxis, including BYD, Hyundai, and Nissian. Nvidia is also partnering with Uber, connecting robo-taxis into Uber's network in select cities.

Olaf joins Jensen on stage

Nvidia GTC keynote stage.

(Image credit: Nvidia)

In a not-at-all-awkward meeting, Olaf from Frozen joins Jensen on stage. The executive is now describing the various AI models used to make Olaf, which is... something. Anyway, Olaf is helping close out the keynote.

That's a wrap, with country song that's probably generated by AI

A video at GTC 2026.

(Image credit: Tom's Hardware)

Well, we're done now, I guess. Nvidia closed out its keynote with an animation of several robots (plus Jensen) sitting around a fire singing a song about the keynote. It's a country song, and probably generated by AI? I don't really know what to say about this one. A rough-talking robot singing about tokens and open-source software wasn't on my bingo card.

  • S58_is_the_goat
    Hey Jensen can ai bring the price of ram down? Thanx bye...
    Reply
  • abufrejoval
    At the heart of this new direction, stands a deep crisis: where for many years the industry thought that the uncanny valley of CGI could be overcome via advances in ray tracing algorithms, Nvidia and others could clearly see that this wasn't going to happen: even at exponential effort for CGI, realism wasn't approaching reality.

    But AI diffusion models delivered awsome results! And that's why some years ago Nvidia decided to change tracks completely, while finding intermediate technologies and products to pay for the transition.

    The idea is to completely replace the algorithmic/mathematical mix of simulation models for game (or in fact any visual) content with an approach that is purely based on machine learning models.

    Perhaps still as an intermediate it might use the same inputs that most game engines deliver to GPUs, the 3D meshes and the textures, but then apply them not in the current algorithmic manner, but mostly as "prompts" for a render that's similar to diffusion and/or how they interpolate frames today.

    If you took the famous teapot as geometry data, I guess it's quite easy to imagine how a "post-Phong-render" or even post-raytrace-render AI-PU might be able to generate something that looks way more naturalistic, even if it has nothing todo with what the designer of the teapot had in mind "exactly".

    The problem might be that than an AI would really need the information that the triangle mesh is supposed to represent a teapot, to do a good job, identifying the teapot via a recogntion model using a simple internal trangle mesh render might work, but ideally you'd want to use the highest level of annoation you can get, which is from the brain of the game designer, if those are still around.

    In short the result looking natural is vastly more important in a game, than the result being authentic or an exact reproduction of the designer's intent (or even the strict adherence to the triangle mesh). And who in Hollywood was ever interested in realism (warts!), when they sold art or illusion?

    Of course modern scene data is slightly more complex than a teapot and re-interpreting a scene for which you have no game developer's abstracted story data, but Nvidia might want to sell a few generations of products and game developers would need to adapt, too.

    If you look at video generation tools today, their abilty to produce short bursts of content that looks incredibly realistic, can be quite good, it's mostly when things get longer or scenes/perspectives change completely where things go off the rails. With NPCs or map data the huge currently unexploited advantage there is that their future actions, position, appearance etc. can be known and thus "prompted" long before they need to be rendered, so errors don't need to accumulate as they do in current video generation models, where the prompt remains static.

    It's with vast distributed multi-player games where predictability can fail, because the shared decisions can't be agreed until real-time has passed, where I see a new wave of jitters and bugs befalling their outputs. PvE could become really rather incredible, eventually, if the money keeps flowing.

    So the biggest benefit may actually be for the content creation industry, where the last human actors may go the way of the dodo or go back to putting out a begging bowl during live performances somewhere where humans can offer them a ball of rice in return for their antics.

    BTW: I have zero insights into what Nvidia actually does. This is pure BI (bio intelligence) hallucinations from hints dropped here and there.

    P.S.: the change of approach was hinted at years ago. But the way I imagined it to go, it would have required changing the game engines to offer the high quality/high abstraction and temporally sorted metadata to the AI-PUs well in advance.

    I've heard zero reports on say Valve adding such functionality to a future release of their engine, so either I'm very wrong, or they managed to pull up a pretty impressive cover-up.
    Reply
  • SkyBill40
    All that... and not a damned thing about consumer space GPUs. Gee, thanks, AI.
    Reply
  • RoLleRKoaSTeR
    Coming soon.
    You will never own anything - House, car, media - ever again. Nor the right to repair or modify equipment you "rent"
    Computer? - nope - it is now a thin client box that you will have to rent and if it needs GPU ability - you will have to rent that, too.
    Reply
  • abufrejoval
    RoLleRKoaSTeR said:
    Coming soon.
    You will never own anything - House, car, media - ever again. Nor the right to repair or modify equipment you "rent"
    Computer? - nope - it is now a thin client box that you will have to rent and if it needs GPU ability - you will have to rent that, too.
    Nonsense, without a job you can't pay rent, neither.
    Reply
  • usertests
    https://i.ytimg.com/vi/I4HmlBRWRjQ/hqdefault.jpg?sqp=-oaymwEXCOADEI4CSFryq4qpAwkIARUAAIhCGAE=&rs=AOn4CLCaZz5mULO6WiusgDDbbwe_Cs0qwA
    abufrejoval said:
    Nonsense, without a job you can't pay rent, neither.
    UBI by Olaf
    SkyBill40 said:
    All that... and not a damned thing about consumer space GPUs. Gee, thanks, AI.
    My hopes of RTX 5050 9GB: Dashed
    Reply
  • abufrejoval
    I felt there was a lot of Cerebras in what they do with Groq.
    Reply
  • usertests
    abufrejoval said:
    I felt there was a lot of Cerebras in what they do with Groq.
    Mentioned in other article: https://www.tomshardware.com/pc-components/gpus/nvidia-groq-3-lpu-and-groq-lpx-racks-join-rubin-platform-at-gtc-sram-packed-accelerator-boosts-every-layer-of-the-ai-model-on-every-token
    The addition of the Groq 3 LPU to the Rubin arsenal could help the platform fend off challengers in the low-latency inference frontier. Cerebras, whose wafer-scale engines fuse massive amounts of SRAM and compute for low-latency inference with advanced models, has frequently needled Nvidia regarding the perceived disadvantages of its GPUs for that purpose, and customers as large as OpenAI have signed up for Cerebras capacity to serve some of their state-of-the-art models with the favorable latency characteristics of that platform.
    Reply
  • abufrejoval
    Let's just say this for Jensen: he is the only one in the techbro field, whose visions keep up scaling with the technology he's enabling. Quite a few struggle to keep up with what he's doing, and he's doing the harder part of pushing.

    It doesn't quite solve the issue of how consumers will be able to consume what he delivers or enables when they can't earn, nor things like the fragility of these technology and data centers to pesty drones strikes: everybody will be hunting for that new "first strike" capability, that leaves the opponent without any ability to continue the fight, no matter what the collaterals: a prequel is playing in Hormuz.
    Reply