Sign in with
Sign up | Sign in

DirectX 11: More Notable Than DirectX 10?

ATI Radeon HD 5870: DirectX 11, Eyefinity, And Serious Speed
By , Fedy Abi-Chahla

I still remember when ATI launched the original Radeon in 2000. Brian Hentschel called me up to ask for my opinion on the chip’s name, and I remember thinking Radeon was horrible. Shows you how well I’d do in PR. That architecture emphasized the gaming experience, with two pixel pipelines and three texture units per pipeline. Though the Radeon had a fairly high texel rate, pixel fill rate was what won the day back then, and the decision to “go pretty” ceded the performance battle to Nvidia.

With R300 in 2002, ATI went the other way, putting its money behind eight pixel pipelines, each with a single texture unit—the focus was on performance—and the bet paid off. ATI smoked the GeForce4 Ti 4600 and fared well enough against an embarrassingly-loud GeForce FX 5800 Ultra.

Performance or experience—which is better? With the Radeon HD 5870, ATI says it’s gunning for both. We’ve already covered the architecture, ATI’s key to delivering performance with Cypress. Now let’s take a closer look at the experience. ATI is relying on three components enabled through hardware here: its Eyefinity technology, Stream, and DirectX 11.

Right off the bat, I think it’s fairly safe to say that for all of the hoopla ATI made about its DirectX 10.1 support, real gamers in the real world never saw a tangible benefit. I’ve played S.T.A.L.K.E.R.; I’ve played H.A.W.X. The experience on a DirectX 10 card versus DirectX 10.1 is not worth mention. And for that matter, I’d also argue that DirectX 10 hasn’t had as profound on impact on gaming experience as prior versions of the API. Why should we believe DirectX 11 is going to be any more prolific than its predecessor?

The reality of the situation is that DirectX 11 probably won’t be as impactful as DirectX 8 or 9, both of which introduced key shading capabilities. But it is seen as the next logical step for ISVs still working with DirectX 9, since it’s a super-set of DirectX 10/10.1 supporting existing hardware, plus DX11 cards. Microsoft has made sure it's easier to code with DX11, so we really are expecting to see a faster up-take of the API than DirectX 10.

New Features

The notable features supported by DX11 are illustrated in the chart below.

Feature
DirectX 10
DirectX 10.1
DirectX 11
Tessellation
-
-
x
Shader Model
-
-
x
DirectCompute 11
-
-
x
DirectCompute 10.1
-
x
x
DirectCompute 10
x
x
x
Multi-Threading
x
x
x
BC6/BC7 Texture Compression
-
-
x


ATI has included tessellation support in its GPUs since 2001. And while I’m not sure how much play those early hardware implementations actually got in the game development world, they’ve helped pave the way for tessellation as it exists today, exposed through a number of different implementations (Catmull-Clark subdivision surface modeling, Bézier patch meshes, n-patches, displacement mapping, and adaptive/continuous tessellation). Of course, the benefits of tessellation are apparent—more polygons mean more detail and hence more realism. And because tessellation is now standardized as a component of DirectX 11, ISVs are more likely to lean on it without the frustration of only supporting one vendor’s hardware. In fact, we saw Rebellion demonstrate tessellation in its upcoming Aliens Vs. Predator title, launching in Q1 of next year.

Perhaps one of the first games to offer DirectX 11-based tessellation, Aliens vs. Predator will emerge in Q1 '10. Perhaps one of the first games to offer DirectX 11-based tessellation, Aliens vs. Predator will emerge in Q1 '10.

DirectX 11 also introduces Shader Model 5.0, which offers developers a more object-oriented approach to coding HLSL. Ideally, this will help motivate ISVs to adopt the new API quicker, since programming becomes cleaner and more efficient (for a more specific example of this in practice, check out Fedy’s DirectX 11 preview).

Gamers with multi-core processors should realize performance gains from DirectX 11-based titles by virtue of threading optimizations made to the API. ATI and Nvidia have been shipping multi-threaded drivers for three years now capable of dispatching commands to the GPU in parallel. According to Nvidia, this was worth anywhere from 10 to 40 percent additional performance back in ’06. But DirectX 11 goes even further, allowing the application, DirectX runtime, and driver to run in separate threads. ATI gives the example of loading textures or compiling shaders in parallel with the main rendering thread. In essence, the threading is much more granular, which, almost ironically, should prove more beneficial to AMD and Intel than ATI and Nvidia. Wait. Yeah. That’s a real win for AMD, isn’t it?

From Allison Klein's Gamefest 2008 presentationFrom Allison Klein's Gamefest 2008 presentation

Improved texture compression is another one of those developer-oriented enhancements that will benefit gamers through greater rendering quality without the expected corresponding performance hit to memory bandwidth. DirectX 11 includes two new block compression formats: BC6 and BC7. BC6 enables up to 6:1 compression of 16-bit high dynamic range textures with hardware decompression support. BC7 delivers up to 3:1 compression of eight-bit textures and normal maps.

Ask a Category Expert

Create a new thread in the Reviews comments forum about this subject

Example: Notebook, Android, SSD hard drive

Display all 214 comments.
This thread is closed for comments
Top Comments
  • 26 Hide
    hispeed120 , September 23, 2009 4:13 AM
    I'm. So. Excited.
  • 23 Hide
    cangelini , September 23, 2009 4:43 AM
    viper666why didn't they thest it against a GTX 295 rather than 280??? its far superior...


    Ran it against a GTX 295 and a 285 and 285s in SLI :) 
  • 22 Hide
    megamanx00 , September 23, 2009 4:48 AM
    O M F G!!!!!!!!!!!!!!!!!!!!!!!!!!!!!

    Just wish the darn thing wasn't so big, but man, what a card! Now I'm thinking about a bigger case :D 
Other Comments
  • 26 Hide
    hispeed120 , September 23, 2009 4:13 AM
    I'm. So. Excited.
  • 9 Hide
    Anonymous , September 23, 2009 4:15 AM
    Can't wait
  • 21 Hide
    crosko42 , September 23, 2009 4:21 AM
    So it looks like 1 is enough for me.. Dont plan on getting a 30 inch monitor any time soon.
  • 20 Hide
    jezza333 , September 23, 2009 4:29 AM
    Looks like the NDA lifted at 11:00PM, as there's a load of reviews now just out. Once again it shows that AMD can produce a seriously killer card...

    Crysis 2 on an x2 of this is exactly what I'm waiting for.
  • 8 Hide
    woostar88 , September 23, 2009 4:38 AM
    This is incredible at the price point.
  • 20 Hide
    tipmen , September 23, 2009 4:40 AM
    wait, wait, before I look can it play cry... HOLY SHIT?!
  • 23 Hide
    cangelini , September 23, 2009 4:43 AM
    viper666why didn't they thest it against a GTX 295 rather than 280??? its far superior...


    Ran it against a GTX 295 and a 285 and 285s in SLI :) 
  • 2 Hide
    Annisman , September 23, 2009 4:44 AM
    I refuse to buy until the 2GB versions come out, not to mention newegg letting you buy more than 1 at a time, paper launch ftl.
  • 15 Hide
    jasperjones , September 23, 2009 4:44 AM
    Thanks for the timely review. I have to say though, some of the technical details are beyond me. It'd be useful if you explained terms such as "VLIW architecture" or "tessellation engine"
  • 22 Hide
    megamanx00 , September 23, 2009 4:48 AM
    O M F G!!!!!!!!!!!!!!!!!!!!!!!!!!!!!

    Just wish the darn thing wasn't so big, but man, what a card! Now I'm thinking about a bigger case :D 
  • 17 Hide
    Annisman , September 23, 2009 4:49 AM
    Oops, who am I kidding ? I just ordered 2 5870's. One Sapphire, and one HIS, seeing as how they limit you to one per customer.
  • 18 Hide
    falchard , September 23, 2009 4:54 AM
    I think most of this review has to do with how many games are optimized for nVidia. The Crytek Engine 2.0 and Source Engine are well known for heavily favoring nVidia architecture yet compose the bulk of the benchmarks. I think the fact ATI can do best in these engines when they have a detect ATI instant nerf its performance speaks measures for the actual card.
  • 14 Hide
    tipmen , September 23, 2009 4:56 AM
    Another thing is that the 5800x2 isn't out yet, now think of two of those bad boys in Crossfire.
  • 6 Hide
    blackbyron , September 23, 2009 4:59 AM
    Not bad for Crysis benchmark. I really want 5870 for my christmas present, but damn I also need to buy a new PSU.
  • 12 Hide
    blackbyron , September 23, 2009 5:02 AM
    In addition, I am impressed that the 5870 has a better power consumption and better gaming performance compare to DX10 cards. If the card is affordable I'd definite buy one.
  • 3 Hide
    cangelini , September 23, 2009 5:10 AM
    jasperjonesThanks for the timely review. I have to say though, some of the technical details are beyond me. It'd be useful if you explained terms such as "VLIW architecture" or "tessellation engine"


    Jasper,
    TBH, the architectural details are secondary to how the card performs. However, if you'd like a better idea of what tessellation can do for you, check out the picture of the Alien on page six!
Display more comments