Page 1:GeForce GTX 580 Goes To Eleven
Page 2:GF110: Nvidia Gives Fermi A Facelift
Page 3:GeForce GTX 580: Similar Dimensions, Improved Design
Page 4:Test Setup And Benchmarks
Page 5:Benchmark Results: 3DMark Vantage
Page 6:Benchmark Results: Metro 2033 (DX11)
Page 7:Benchmark Results: Lost Planet 2 (DX11)
Page 8:Benchmark Results: Aliens Vs. Predator (DX11)
Page 9:Benchmark Results: Battlefield: Bad Company 2 (DX11)
Page 10:Benchmark Results: Civilization 5 (DX11)
Page 11:Benchmark Results: DiRT 2 (DX11)
Page 12:Benchmark Results: Just Cause 2 (DX11)
Page 13:Benchmark Results: Adobe Premiere Pro CS5
Page 14:Benchmark Results: SLI
Page 15:Benchmark Results: Power And Heat
GeForce GTX 580: Similar Dimensions, Improved Design
Change For The Better
Because Nvidia chose to spend the thermal budget it freed up on performance (more shaders, higher clocks), the GeForce GTX 580’s TDP is very similar to the GeForce GTX 480 (244 W versus 250 W). As a result, you might expect the card to run into similar heat and noise issues as its predecessor. Fortunately, as the architects were reworking the GPU itself, another team tackled 480’s shortcomings.
Gone are the exposed heatsink, the protruding heat pipes, and the noisy fan.
A vapor chamber-based cooler enables those changes. Heatpipes disappear completely, as the copper vapor chamber has its own evaporation/condensation cycle to more effectively transfer heat to a fin array. Nvidia still uses a blower-style fan to move air across the heatsink and out the rear I/O panel, but it’s reinforced to drop pitch. Fan control migrates into the GPU itself, facilitating a more gradual (less noticeable) ramp up and down in response to activity.
The overall result is a completely contained card that’s still hot, but nowhere near as painful to touch after a bout on the bench system. It’s remarkably quiet at idle, and much more pleasant under load. Perhaps I should repeat that for emphasis. Excessive noise was one of the strongest reasons to avoid GeForce GTX 480, and it is effectively dealt with on the GTX 580.
Aside from those tweaks, the GeForce GTX 580 is as long as 480 (10.5”). It still requires one eight-pin and one six-pin auxiliary PCIe power connector (Nvidia additionally recommends at least a 600 W power supply).
And it’s still limited to two simultaneous display outputs via two dual-link DVI connectors and a mini-HDMI port. With AMD now able to support up to four simultaneous displays using a combination of DisplayPort, DVI, and HDMI (from a mid-range lineup, no less), Nvidia really needs to address its two-output limitation soon. That's a practical advantage on AMD's side that goes unanswered unless you buy two GeForce cards. Naturally, the GTX 580 is still a dual-slot board, too.
Protection-oriented circuitry is quickly changing the way power, temperature, and noise measurements can be made.
Back when AMD launched the Cypress GPU, it introduced a feedback mechanism to read the state of the card’s voltage regulator. If it overvolted (even prior to the GPU overheating), the ASIC would throttle down to drop power draw. This was in addition to the thermal protection scheme that’d simply keep the chip from getting too hot. As a result, it was entirely possible to run a constant-load application like FurMark and see unrealistically low power numbers without the tell-tale high GPU temps to indicate that something was wrong. Fortunately, I figured out a combination of settings that’d still tax the GPU without triggering either of AMD’s mechanisms.
Power-monitoring circuitry added to GeForce GTX 580
Nvidia’s a little late to this party, but it’s implementing something similar on GeForce GTX 580. The company claims its GPUs have had thermal monitoring for many years (though that doesn’t explain why its 196-series driver sent a number of G92-based boards to an early grave earlier this year). Now, the current and voltage of each 12 V rail is being monitored as well. Should power levels exceed spec, the GPU throttles performance. More on this once we get to power testing.
Update: Multi-Display Testing
I had a couple of requests to check back on issues that originally plagued the GeForce GTX 480, one of those being a dramatic increase in power consumption and heat with two displays attached.
According to Nvidia, it rectified the out-of-control increases that were being reported in the GeForce 256-series driver released earlier this year. So long as you're using two monitors with the same resolution and timing settings, you're supposedly safe. In an effort to double-check/verify, I attached a pair of Dell P2210H displays to a GeForce GTX 580 and charted out temps and power use:
|One Display (1920x1080), Idle||190 W||40 deg. C |
|Two Displays (1920x1080), Idle||192 W||45 deg. C|
|Two Displays (1 x 1920x1080, 1 x 1280x1024), Idle||255 W||56 deg. C|
Power consumption doesn't increase much when you attach a second display running at the same resolution and timings, but the temperature does increase by five degrees.
Swapping over to a display running a different resolution, however, continues to have a profound effect on power and temperatures (Nvidia does not deny this). The jump from 192 W to 255 W and 45 degrees to 56 degrees is significant. The good news is that if you're using the same screen, the latest drivers minimize the impact of utilizing both of the GeForce GTX 580's display outputs.
- GeForce GTX 580 Goes To Eleven
- GF110: Nvidia Gives Fermi A Facelift
- GeForce GTX 580: Similar Dimensions, Improved Design
- Test Setup And Benchmarks
- Benchmark Results: 3DMark Vantage
- Benchmark Results: Metro 2033 (DX11)
- Benchmark Results: Lost Planet 2 (DX11)
- Benchmark Results: Aliens Vs. Predator (DX11)
- Benchmark Results: Battlefield: Bad Company 2 (DX11)
- Benchmark Results: Civilization 5 (DX11)
- Benchmark Results: DiRT 2 (DX11)
- Benchmark Results: Just Cause 2 (DX11)
- Benchmark Results: Adobe Premiere Pro CS5
- Benchmark Results: SLI
- Benchmark Results: Power And Heat