Opinion: AMD, Intel, And Nvidia In The Next Ten Years

Have 3D Graphics Reached The Point Of Diminishing Returns?

Anyone who's been following PC gaming graphics knows that Nvidia has been recycling the G92 core for several product cycles now. Introduced in 2007 with the GeForce 8800 GT, the G92 core has been reincarnated as the GeForce 9600, 9800, GTS 150, and the GTS 250. Nvidia has been able to "get away" with incrementally dropping the price over the last two years, because at medium quality settings, these G92-based cards can still handle games like Far Cry 2 at 30 frames per second at lower resolutions.

On the other extreme, the AMD Radeon HD 5970 is so fast that even with a 30" monitor and everything set to ultra-high quality, you're still seeing greater than 100 FPS in many games. Pause for a moment and think about what AMD’s Eyefinity technology signifies: AMD's performance is so high that it has to come up with creative (and almost frivolous) ways to utilize that power by rendering two or three times that number of pixels. 

This scenario would not have been possible several years ago. When people were buying the GeForce 6800-series flagships, it was the entry-point for 60 fps in games like Doom 3 with Ultra High Quality. We weren't even close to talking about insane resolutions and multiple monitors yet.

Simply put, software development has not been moving as fast as hardware growth. While hardware manufacturers have to make faster and faster products to stay in business, software developers have to sell more and more games. Developing games that only run at 30 FPS on a Radeon HD 5970 requires an enormous budget to develop the art assets, and offers a very limited potential audience.  

The major threat to the hardware industry, ironically enough, comes from the software industry. Software developers must continue to develop products that take advantage of the latest and greatest hardware. Their titles have to be compelling enough in game play to get people to invest in more powerful computing hardware.

  • anamaniac
    Alan DangAnd games will look pretty sweet, too. At least, that’s the way I see it.After several pages of technology mumbo jumbo jargon, that was a perfect closing statement. =)

    Wicked article Alan. Sounds like you've had an interesting last decade indeed.
    I'm hoping we all get to see another decade of constant change and improvement to technology as we know it.

    Also interesting is that you almost seemed to be attacking every company, you still managed to remain neutral.
    Everyone has benefits and flaws, nice to see you mentioned them both for everybody.

    Here's to another 10 years of success everyone!
    Reply
  • False_Dmitry_II
    I want to read this again in 10 years just to see the results...
    Reply
  • " Simply put, software development has not been moving as fast as hardware growth. While hardware manufacturers have to make faster and faster products to stay in business, software developers have to sell more and more games"

    Hardware is moving so fast and game developers just cant keep pace with it.
    Reply
  • Ikke_Niels
    What I miss in the article is the following (well it's partly told):

    I am allready suspecting a long time that the videocards are gonna surpass the CPU's.
    You allready see it atm, videocards get cheaper, CPU's on the other hand keep going pricer for the relative performance.

    In the past I had the problem with upgrading my videocard, but with that pushing my CPU to the limit and thus not using the full potential of the videocard.

    In my view we're on that point again: you buy a system and if you upgrade your videocard after a year/year-and-a-half your mostlikely pushing your CPU to the limits, at least in the high-end part of the market.

    Ofcourse in the lower regions these problems are smaller but still, it "might" happen sooner then we think especially if the NVidia design is as astonishing as they say and on the same time the major development of cpu's slowly break up.


    Reply
  • sarsoft
    Nice article. Good read....
    Reply
  • lashton
    one of the most interesting and informativfe articles from toms hardware, what about another story about the smaller players, like Intel Atom and VILW chips and so on
    Reply
  • JeanLuc
    Out of all 3 companies Nvidia is the one that's facing the more threats. It may have a lead in the GPGPU arena but that's rather a niche market compared to consumer entertainment wouldn't you say? Nvidia are also facing problems at the low end of market with Intel now supplying integrated video on their CPU's which makes the need for low end video cards practically redundant and no doubt AMD will be supplying a smiler product with Fusion at some point in the near future.
    Reply
  • jontseng
    This means that we haven’t reached the plateau in "subjective experience" either. Newer and more powerful GPUs will continue to be produced as software titles with more complex graphics are created. Only when this plateau is reached will sales of dedicated graphics chips begin to decline.
    I'm surprised that you've completely missed the console factor.

    The reason why devs are not coding newer and more powerful games is nothing to do with budgetary constraints or lack thereof. It is because they are coding for an XBox360 / PS3 baseline hardware spec that is stuck somewhere in the GeForce 7800 era. Remember only 13% of COD:MW2 units were PC (and probably less as a % sales given PC ASPs are lower).

    So your logic is flawed, or rather you have the wrong end of the stick. Because software titles with more complex graphics are not being created (because of the console baseline), newer and more powerful GPUs will not continue to produced.

    Or to put it in more practical terms, because the most graphically demanding title you can possibly get is now three years old (Crysis), then NVidia has been happy to churn out G92 respins based on a 2006 spec.

    Until we next generation of consoles comes through there is zero commercial incentive for a developer to build a AAA title which exploits the 13% of the market that has PCs (or the even smaller bit of that has a modern graphics card). Which means you don't get phat new GPUs, QED.

    And the problem is the console cycle seems to be elongating...

    J
    Reply
  • Swindez95
    I agree with jontseng above ^. I've already made a point of this a couple of times. We will not see an increase in graphics intensity until the next generation of consoles come out simply because consoles is where the majority of games sales are. And as stated above developers are simply coding games and graphics for use on much older and less powerful hardware than the PC has available to it currently due to these last generation consoles still being the most popular venue for consumers.
    Reply
  • Swindez95
    Oh, and very good article btw, definitely enjoyed reading it!
    Reply