Tom's Hardware Graphics Charts: Performance In 2014
Two years and two graphics card generations have passed since the last major update to our famous graphics card performance charts. It's time to get them back up to speed. We introduce modern benchmarks, new measurement equipment, and fresh methodology.
Metro: Last Light And Thief
Metro: Last Light
Is it better to create your own benchmark, or use a game's built-in test designed to deliberately push graphics hardware? In this case, Metro: Last Light includes a tool for dialing-in settings and creating repetition. It's a worst-case example of what your GPU will have to endure when you play. And, if you already own Metro, it's easy to replicate the options we picked and compare your machine's performance.
The following video shows one of the four benchmark runs we execute. The first loop heats the GPU being tested, while results from the last three are averaged together.
The 4A engine pushes almost every graphics card to its limit, so its inclusion is intended to represent some of the lowest performance you'll see from any given board.
Metro: Last Light | |
---|---|
Run 1 | 1920x1080 (1080p)API: DirectX 11Quality: Very HighAF: 16xMotion Blur: NormalTessellation: NormalSSAA: No |
Run 2 | 3840x2160 (2160p)API: DirectX 11Quality: HighAF: 16xMotion Blur: LowTessellation: NormalSSAA: No |
Loops | Four per resolution; three used for evaluation |
Thief
Thief is demanding in its own right. It also includes a built-in benchmark, which gives you an open invitation to do some comparative testing at home. That metric is quite memory-heavy and it'll punish any graphics card without enough on-board RAM to handle the resolution and settings you pick.
The test is short enough that we're able to run it three times back-to-back. Again, the first iteration heats each GPU, while the second two are averaged. The video shows the benchmark sequence we use for our chart results.
And here are the settings in a table:
Thief | |
---|---|
Run 1 | 1920x1080 (1080p)Full-screen Mode (Exclusive)V-sync: OffEngine: 64 BitPreset: Very High |
Run 2 | 3840x2160 (2160p)Full-screen Mode (Exclusive)V-sync: OffEngine: 64 BitPreset: Normal |
Loops | Three per resolution; two used for evaluation |
Stay On the Cutting Edge: Get the Tom's Hardware Newsletter
Get Tom's Hardware's best news and in-depth reviews, straight to your inbox.
Current page: Metro: Last Light And Thief
Prev Page 3DMark Fire Strike And Unigine Heaven Next Page DiRT 3 And BioShock Infinite-
blackmagnum Thank you Tom's team for updating the charts. You're my goto when I'm upgrading my rigs. I'll be waiting... Bring on yesterday's gems.Reply -
tomfreak First thing Tom need is to bench how PCIE 2.0 8x vs 16x perform on a modern top end GPU. Since 290X are passing the bandwidth from crossfire bridge to PCIE, may be is time to check them again? As I recall AMD do not recommend putting 290x XDMA crossfire on PCIE 2.0 8x. Please check this outReply -
cypeq First it's great to see new charts.Reply
I was never a fan of this style of benchmarking. It sure gives clean graph of gpu capabilities which we always needed. I would love to see new bottleneck analysis. Or at least parallel test done on midrange PC.
Everyone should keep mind that these charts represent performance of <1% PC builds out there.
13278215 said:First thing Tom need is to bench how PCIE 2.0 8x vs 16x perform on a modern top end GPU. Since 290X are passing the bandwidth from crossfire bridge to PCIE, may be is time to check them again? As I recall AMD do not recommend putting 290x XDMA crossfire on PCIE 2.0 8x. Please check this out
If I recall correctly we are at this moment at the edge of PCI 2.0 x8 which = PCI 1.0 x16 . Next or following gen will finally outdate PCI 1.0 in single and PCI 2.0 in dual GPU configs as there will finally be noticeable bottle necks. -
mitcoes16 Any Steam OS or GNU/Linux benchmarks?Reply
It would be nice to add any opengl crossplattform game as any ioquake based one or something more modern and test it under MS WOS and under GNU / Linux
Better if it is future Steam OS to let us know the performance at the same game under MS WOS and under GNU/Linux.
Also it would be nice to test at MS WOS with and without antivirus, perhaps avast that is free or any other of your preference.
Last but not least, in opengl or in directx there are version changes and being able to split cards generations by opengl / directx version support would help as a current price / performance index based in your sponsored links prices. -
mitcoes16 No 720p tests?Reply
720p ( 1280x720 píxels = 921.600 píxels) is half 1080p more or less
1080p (1920\00d71080 píxels = 2.073.600 pixels)
And when a game is very demanding or you prefer to play with better graphics playing at 720p is a great option
Of course,latest best GPUs would be able to play at 4k and full graphics, but when we read the benchmarks we want to know also if our actual card CAN play at 720p (1k) or what the best ones can do at 1k to be able to compare
Also even it is not a standard or accurate, for benchmarking purposes calling 720p (1k) 1080p (2k) and 2160p (4K) wouldbeeasier to understand in a fast sight than UHD FHD and HDR, that can be used too UHD (4k) FHD (2k) HDR (1k) -
InvalidError
720p does not stress most reasonably decent GPUs much and how many people would drop resolution to 720p these days with all the re-scaling artifacts that might add? In most cases, it would make more sense to stick with native resolution and tweak some of the more GPU/memory-intensive settings down a notch or two - at least I know I greatly prefer cleaner images over "details" that get blurred by the lower resolution and re-scaling that further distorts it.13278758 said:No 720p tests?
Considering how you can get 1080p displays for $100, I would call standardizing the GPU chart on 1080p fair enough: the people who can only afford a $100 display won't care much about enabling every bell and whistle and the people who want to max everything out likely won't be playing on $100 displays and $100 GPUs either. -
2Be_or_Not2Be I really like to see the charts on how much noise a video card's cooling fans make. That makes more of a difference to me as limiting something distracting that I hear every time I game versus getting a louder card with 10 fps more.Reply
I also like seeing how current cards stack up performance-wise to previous generations. That really helps when you're deciding whether to upgrade or not.