The Nvidia GeForce RTX 3090 is now confirmed as the next halo graphics card from Team Green, and Jensen has spilled the beans (most of them, anyway) on specs and performance. If you want the best performance from Nvidia's Ampere architecture, get ready to take out a small loan, because the king of the GPU hierarchy and the best graphics card ('best' as in 'fastest') won't come cheap. The RTX 3090 sets a new high-bar for single-GPU pricing at $1,499, not counting Nvidia's Titan series that it's apparently meant to replace. Here's everything we know about the GeForce RTX 3090.
We've covered the high-level view of Nvidia's new Ampere GPUs elsewhere, and you can read about the GeForce RTX 3080 and GeForce RTX 3070 in their own dedicated articles. The focus here is on the RTX 3090. After months of speculation and waiting, we finally have the hard details. It's big, quite literally. Nvidia's RTX 3090 Founders Edition sports a triple-slot cooler and has a 350W TDP. You might need a PSU, case, and CPU upgrade to make the most of this bad boy.
The GeForce RTX 3090 is the first 90-series suffix we've seen from Nvidia since the GTX 690 back in 2012. That was a dual-GPU variant of the GTX 680, back when multi-GPU was a thing. Which, technically it still is, but support has been seriously lacking of late. Regardless, the RTX 3090 is the only Ampere GeForce GPU that has NVLink support this round, just in case you have $3,000 sitting around. (Don't do it!) But let's hit the specs.
Nvidia GeForce RTX 3090 At A Glance:
- 24GB GDDR6X at 19.5Gbps
- 10496 CUDA cores and 35.7 TFLOPS of FP32 compute
- Samsung 8N manufacturing process
- 1.9 times more efficient than Turing
- Release Date: September 24, 2020
- Price: $1,499
|Graphics Card||GeForce RTX 3090|
|Process (nm)||Samsung 8N|
|Die Size (mm^2)||628.4|
|Boost Clock (MHz)||1695|
|VRAM Speed (Gbps)||19.5|
|Tensor TFLOPS FP16 (Sparsity)||142 (285)|
|Launch Date||September 24, 2020|
Nvidia GeForce RTX 3090 Specifications
Holy TFLOPS, Batman! No, seriously: Wow! We heard various rumors. We heard Nvidia might double the number of shader cores per SM. What we didn't expect was double the FP32 shaders while still packing 82 SMs. And Nvidia could have theoretically gone even bigger (the GA100 is an 826mm square chip, where GA102 is apparently only around 627mm square). Still, the resulting 36 TFLOPS of compute is going to be a massive boost to performance ... provided the rest of your PC can keep up.
Raw compute power is 150% more than the RTX 2080 Ti, for both the CUDA cores and the Tensor cores. As in, on paper the RTX 3090 is 2.5 times as fast as the previous king. Actually, maybe that's not fair — it should be compared with the Titan RTX, right? Then it's only 2.2 times as fast, plus it costs $1,000 less.
Except ... it's not that fast. See, the Turing architecture added a second pipeline for INT workloads, and Nvidia says games use a ratio of about 65:35 for FP32:INT. Ampere takes the INT pipeline and turns it into an FP32 or INT pipeline. That means perhaps two thirds of the secondary pipeline will be kept busy with INT operations, while the dedicated FP32 pipeline plugs along at 'only' 17.8 TFLOPS. Or combined, the GPU will behave somewhat like a 23.1 TFLOPS GPU with 12.5 TOPS of INT performance. Still, it's plenty fast.
I'm a bit sad that the GDDR6X memory 'only' clocks in at 19.5Gbps, and I'm in need of a GPU hat to eat (chocolate, please!), but we're still looking at 24GB of memory and 936 GBps of bandwidth. That's a 52% increase relative to the RTX 2080 Ti, and Nvidia likely has some architectural improvements to that it makes better effective use of that bandwidth. Note that GDDR6/GDDR6X have a half-width mode where each chip can run on a 16-bit interface, which is how you get 24 chips with a 384-bit interface and not a 768-bit interface.
Finally, ray tracing performance is 69 TFLOPS of RT computation. Nvidia rated the previous Turing GPUs in gigarays per second, but that was misleading, so we're now getting RT-TFLOPS. The RTX 2080 Ti incidentally had 34 TFLOPS of RT prowess. Again, Nvidia is looking at more than double the computational power on all the core metrics, and a bit more than 50% more memory bandwidth.
Meet the Nvidia GeForce RTX 3090
The GeForce RTX 3090 isn't just the most expensive GeForce card to date; it's the largest graphics card Nvidia has sold. We've seen various third party designs push the limits of good sense (in a good way, provided you have a large PC), but Nvidia has previously limited its designs to 2-slot solutions. No more! The RTX 3090 is a triple-slot card, measuring 12.3-inches in length and 5.4-inches in height.
It's a monster! And I love it. Be still, my heart! Forgive techno-lust, but this is definitely an exciting GPU. Soon, it will be here. My precious... Ahem.
As listed above, the RTX 3090 also sports a 350W TDP (or TGP if you prefer, which is power to the entire GPU and memory). To help cope with the added thermal output, Nvidia has significantly altered the cooling design compared to previous generation GPUs.
The above image is from Nvidia's RTX 3080, but the GeForce RTX 3090 is the same fundamental design — only bigger. We're not positive, but it looks like the 3090 will have a bigger fan (120mm?) to go along with the wider heatsink. The size of the PCB meanwhile is smaller than the previous generation cards, so the extra size really is all about cooling.
It's not too surprising to see Nvidia take this approach. Even though the Turing architecture was very efficient overall, it still ran into power and thermal limits on the fastest models (RTX 2080 Super and above). The only way around that is to increase the TDP, and that meant improving the cooling capabilities as the top RTX 20-series Founders Edition graphics cards could get quite hot.
Part of the redesign also involved moving from dual 8-pin power connectors to a single 12-pin connector, at least for Nvidia's reference design. Third party cards appear to be sticking with dual 8-pin or even triple 8-pin connectors, and the 12-pin cable doesn't necessarily deliver more power. It's just a more compact connector, rotated 90 degrees to free up even more board space.
How will the new design fare against third party cards? We're certainly interested to find out. Will it be quieter, or lower temperatures, or both? Check back in a few weeks and we'll have the details.
Nvidia GeForce RTX 3090 Features
Nvidia's GeForce RTX 3090 will make use of Micron's GDDR6X memory. This GPU and the RTX 3080 are the only ones slated to use GDDR6X for now, and Nvidia had to further improve the signal delivery to boost speeds. There's also new EDR tech for the RAM — Error Detection and Retry. Basically, if the RAM detects an error in transmission, it will retry until it succeeds (or crashes?). This will change memory overclocking somewhat, where at some point you'll increase clocks but end up with lower performance due to EDR.
HDMI 2.1 makes its debut in a graphics card, but the three DisplayPort connectors remain stuck at 1.4a. Both standards can drive an 8K display, but where HDMI 2.1 can do 8K120 via DSC, DisplayPort 1.4a requires DSC just to get to 8K60.
Nvidia has also added PCIe Gen4 support to its Ampere GPU. It's worth pointing out that this probably won't matter much for gaming performance, as the large 24GB of VRAM means there should be less data going back and forth over the PCIe bus. The other problem of course is that the fastest gaming CPUs still come from Intel, and Intel doesn't have a desktop PCIe Gen4 solution yet. That will come with next year's Rocket Lake processors, which will yet again use Intel's 14nm++(++) process. Intel's Alder Lake will also support PCIe Gen4 and will be the first SuperFIN (or post-SuperFIN) desktop CPU from Intel.
Does that mean AMD's X570 platform with a Ryzen 9 3900X is the better choice, since you get Gen4 support? Almost certainly not. We recently ran a full suite of benchmarks on ten GPUs and compared the performance of the Core i9-9900K vs. Ryzen 9 3900X. There were a few edge cases where the 3900X came out ahead, but it was more like a tie. For the RX 5700 XT at least, having a faster PCIe interface didn't appear to matter. Hopefully Zen 3 can further close the CPU gap, but that's a topic for another day.
Nvidia GeForce RTX 3090 Performance
What's the net result of all these improvements? Citius, Altius, Fortius. Or in Nvidia's case: Bigger, faster, better. There's no doubt the RTX 3090 will emerge as Nvidia's fastest GPU ever, and we'd be extremely surprised to see AMD's Big Navi match it. It's not impossible, but it's extremely unlikely. But then it's also $1,500, which is enough to build a complete gaming PC with a GeForce RTX 3070.
How much faster will RTX 3090 be in practice, though? That's the tougher question. We've tested a lot of games and graphics cards over the years, and just in 2020 we've encountered several games where the RTX 2080 Ti didn't come in first place until at least 1080p ultra, and sometimes 1440p ultra. In short, the CPU was a bottleneck, and with significantly more GPU performance, it's going to be an even bigger bottleneck with the RTX 3090.
First piece of advice, then: Don't even bother with the RTX 3090 if you're not using a 4K display (or perhaps 3440x1440 ultrawide). There will be games where it's still the fastest GPU even at 1080p, particularly if the games use ray tracing, but I just can't see 1080p being a good fit for this GPU. If you're playing esports games like CSGO, even an RTX 3070 should suffice for 360 fps. If you're not, then you won't get above 200 fps even at minimum settings in many games.
Nvidia provided the above comparison of ray tracing performance, but it didn't state whether this was for RTX 3090 vs. RTX 2080 Ti, or some other comparison like RTX 3080 vs. RTX 2080. Regardless, even at 4K, in most cases the new Ampere GPU isn't able to double the performance of Turing. Maybe there are other factors, but we wager the CPU is holding the graphics card back at least a little.
If you do have a high-end or one of the best gaming monitors, however, and you want to run with all the bells and whistles? The GeForce RTX 3090 is your best chance at maintaining silky smooth framerates. That will be even more important as the next generation consoles launch and we start to see games use more ray tracing effects.
Besides Cyberpunk 2077, which will use ray tracing for shadows, reflections, ambient occlusion, and diffuse illumination (plus DLSS 2.0), Call of Duty Black Ops Cold War will similarly support dynamic lights, shadows, ambient occlusion, and DLSS 2.0. The Fortnite RTX patch also adds multiple ray tracing effects. And the RTX 3090 is beefy enough that you can probably even use all of those in multiplayer without tanking your framerates.
In short, while the first generation of ray tracing enhanced games wasn't necessarily a great showcase for the technology (Control and perhaps Minecraft RTX being exceptions), it's not going anywhere. Nearly every movie these days makes use of ray tracing. It might be in its infancy as far as gaming is concerned, but over the coming years we expect to see more games pushing better effects. Just don't be surprised when RTX 4090 or RTX 5090 show up in the coming years and make even the RTX 3090 look anemic.
Nvidia GeForce RTX 3090: The Bottom Line
The biggest unknowns have now been answered. The GeForce RTX 3090 is a graphical tour de force, with a massive 10496 GPU cores delivering up to 36 TFLOPS of compute performance. It will officially launch on September 24, priced at $1,500. Don't be surprised if it sells out, and don't be surprised if many custom designs end up costing far more than $1,500.
That's actually the good part about GeForce RTX 3090 branding instead of Titan branding. Nvidia always did their Titan cards in-house, and third party designs weren't allowed. The result was that the cards cost an arm and a leg (and a kidney as well), for a relatively small increase in performance. This is still an extremely expensive graphics card, of course, but it's $1,000 less than the Titan RTX.
Will the RTX 3090 be worth the price? Probably not, at least in terms of fps per dollar spent. You certainly don't need a $1,500 graphics card to enjoy playing games, and you could buy three RTX 3070 cards for the same asking price. It's arguably more about bragging rights and being the fastest GPU on the planet, price be damned.
If you've got deep pockets, by all means, pre-order an RTX 3090 and get ready to rub it in your friends' noses. Why not order two plus an NVLink connector for good measure? You should seriously consider tossing in an 8K display for good measure. The rest of us will wait to see what the real-world gaming performance looks like come September 24.