The implementation of GPU Boost does not preclude overclocking. But because you can’t disable GPU Boost like you might with Intel’s Turbo Boost, you have to operate within the technology’s parameters.
For example, overclocking is now achieved through an offset. You can easily push the base 3D clock up 100, 150, or even 200 MHz. However, if a game was already TDP-constrained at the default clock, it won’t run any faster. In apps that weren’t hitting the GTX 680’s thermal limit before, the offset pushes the performance curve closer to the ceiling.
Because GPU Boost was designed to balance clock rate and voltage with thermal design power in mind, though, overclocking is really made most effective by adjusting the board’s power target upward as well. EVGA’s Precision X tweaking tool includes built-in sliders for both the power target and the GPU clock offset.
Although GeForce GTX 680’s TDP is 195 W, Nvidia says the card’s typical board power is closer to 170 W. So, increasing the power slider actually moves this number higher. At +32%, Precision X’s highest setting is designed to get you right up to the 225 W limit of what two six-pin power connectors and a PCI Express slot are specified to deliver.
Using Crysis 2 as our very-consistent test case, we can measure the impact of each different alteration and its effect on performance.
First, we launch a single run of the Central Park level at 1920x1080 in DirectX 11 mode, without anti-aliasing. We get a 72.3 FPS result, and we observe GPU Boost pushing the GeForce GTX 680 between 1071 and 1124 MHz during the run (up from the 1006 MHz base).
The top chart shows that we’re bouncing around the upper end of GK104’s power ceiling. So, we increase the target board power by 15%. The result is a small jump to 74.2 FPS, along with clocks that vacillate between 1145 and 1197 MHz.
Figuring the power target boost likely freed up some thermal headroom, we then increase the offset by 100 MHz, which enables even better performance—76.1 FPS. This time, however, we get a constant 1215 MHz. Nvidia says this is basically as fast as the card will go given our workload and the power limit.
So why not up the target power again? At 130% (basically, the interface’s 225 W specification), performance actually drops to 75.6 FPS, and the graph over time shows a constant 1202 MHz. We expected more performance, not less. What gives? This is where folks are going to find a problem with GPU Boost. Because outcome is dependent on factors continually being monitored, performance does change over time. As a GPU heats up, current leakage increases. And as that happens, variables like frequency and voltage are brought down to counter a vicious cycle.
The effect is similar to heat soak in an engine. If you’re on a dynamometer doing back to back pulls, you expect to see a drop in horsepower if you don’t wait long enough between runs. Similarly, it’s easy to get consistently-high numbers after a few minute-long benchmarks. But if you’re gaming for hours, GPU Boost cannot be as effective.
Our attempt to push a 200 MHz offset demonstrates that, even though this technology tries to keep you at the highest frequency under a given power ceiling, increasing both limits still makes it easy to exceed the board’s potential and seize up.
Sliding back a bit to a 150 MHz offset gives us stability, but performance isn’t any better than the 100 MHz setting. No doubt, it’ll take more tinkering to find the right overclock with GPU Boost in the mix and always on.
- GeForce GTX 680: The Card And Cooling
- GK104: The Chip And Architecture
- GPU Boost: Graphics Afterburners
- Overclocking: I Want More Than GPU Boost
- PCI Express 3.0 And Adaptive V-Sync
- Hardware Setup And Benchmarks
- Benchmark Results: 3DMark 11 (DX 11)
- Benchmark Results: Battlefield 3 (DX 11)
- Benchmark Results: Crysis 2 (DX 9/DX 11)
- Benchmark Results: The Elder Scrolls V: Skyrim (DX 9)
- Benchmark Results: DiRT 3 (DX 11)
- Benchmark Results: World Of Warcraft: Cataclysm (DX 11)
- Benchmark Results: Metro 2033 (DX 11)
- Benchmark Results: Sandra 2012
- Benchmark Results: Compute Performance In LuxMark 2.0
- Benchmark Results: NVEnc And MediaEspresso 6.5
- Temperature And Noise
- Power Consumption
- Performance Per Watt: The Index
- GeForce GTX 680: The Hunter Scores A Kill


Now we just need prices to start dropping, although significant drops will probably not come until the GK110 is released
Now we just need prices to start dropping, although significant drops will probably not come until the GK110 is released
Good going Nvidia...
Sigh...
WoW has had DX11 for quite a long time now. Also, go play in a 25 man raid with every detail setting on ultra with 8xAA and 16x AAF and tell me WoW is not taxing on a PC.
...oh, wait.
For everyone suggesting that nVidia will release another true "flagship" beyond the 680, I think you are spot on, IF AMD gives them a reason to. There's no reason to push it at the moment as they already hold the crown. If, on the other hand, AMD goes out and makes a 7980, or 79070 SE card with higher clocks (more like what the 7970 can achieve when properly overclocked), I definitely see nVidia stepping their game up a bit.
Either way, it's awesome to see both AMD and now nVidia taking power consumption into consideration. I'm tired of my computer room feeling like a toaster after an all nighter.
He means waiting for the GK110, that will be a more of a compute card while this GK104 is more equiped towards gaming.