Sign in with
Sign up | Sign in

3D LightBoost, On-Board Memory, Standards, And 4K

G-Sync Technology Preview: Quite Literally A Game Changer
By , Chris Angelini

As we were going through Nvidia's press material, we found ourselves asking a number of questions about G-Sync as a technology today, along with its role in the future. During a recent trip to the company's headquarters in Santa Clara, we were able to get some answers.

G-Sync And 3D LightBoost

The first thing we noticed was that Nvidia was sending out that Asus VG248QE monitor, modified to support G-Sync. That monitor also supports what Nvidia currently calls 3D LightBoost technology, which was originally introduced to improve brightness in 3D displays, but has long been unofficially used in 2D mode as well, using its panel-pulsing backlight to reduce the ghosting (or motion blur) artifact we mentioned on page one. Naturally, we wanted to know if that could be used with G-Sync.

Nvidia answered that no, although using both technologies at the same time is what you'd want ideally, today, strobing the backlight at a variable refresh currently results in flicker and brightness issues. Solving them is incredibly complex, since you have to adjust luminance and keep track of pulses. As a result, you currently have to choose between the two technologies, although the company is working on a way to use them together in the future. 

The G-Sync Module's On-Board Memory

As we already know, G-Sync eliminates the incremental input lag associated with V-sync, since there's no longer a need to wait for the panel to scan. However, we noticed that the G-Sync module has on-board memory. Could the module be buffering frames itself? If so, how much time would it take for a frame to make its way through the new pipeline? 

According to Nvidia, frames are never buffered in the module's memory. As data comes in, it's displayed on the screen, and the memory does perform several other roles. However, the processing time related to G-Sync is way less than one millisecond. In fact, this is roughly the same latency encountered with V-sync off, and is related to the game, graphics driver, the mouse, and so on.

Will G-Sync Ever Be Standardized?

This came up during a recent AMA with AMD when a reader wanted to get the company's reaction to G-Sync. However, we also wanted to follow up with Nvidia directly to see if the company had any plans to push its technology as an industry standard. In theory, it could propose G-Sync as an update to the DisplayPort standard, exposing variable refresh rates. Nvidia is a member of VESA, the industry's main standard-setting board, after all. 

Simply, there are no plans to introduce a new spec to DisplayPort, HDMI, or DVI. G-Sync is already working with DisplayPort 1.2, meaning there's no need for a standards change. 

As mentioned, Nvidia is working to make G-Sync compatible with what it currently refers to as 3D LightBoost (and will be called something else soon). It's also trying to bring the module's cost down to make G-Sync more accessible.

G-Sync At Ultra HD Resolutions

Nvidia's online FAQ promises G-Sync-capable monitors with resolutions as high as 3840x2160. However, the Asus model we're previewing today maxes out at 1920x1080. Currently, the only Ultra HD monitors employ STMicro's Athena controller, which uses two scalers to create a tiled display. We were curious, then: does the G-Sync module support an MST configuration?

In truth, we'll be waiting a while before 4K displays show up with variable refresh rates. Today, there is no single scaler able to support 4K resolutions, and the soonest that's expected to arrive  is Q1 of 2014, with monitors including it showing up in Q2. Because the G-Sync module replaces the scaler, compatible panels would only start surfacing sometime after that point. Fortunately, the module does natively support Ultra HD.

Keeping Up: What Happens Under 30 Hz?

The variable refresh enabled by G-Sync works down to 30 Hz. The explanation for this is that, at very low refresh rates, an image on an LCD starts to decay after a while and you end up with visual artifacts. If your source drops below 30 FPS, the module knows to refresh the panel automatically, avoiding those issues. That might mean displaying the same image more than once, but the threshold is set to 30 Hz to maintain the best-looking experience possible.

Ask a Category Expert

Create a new thread in the Reviews comments forum about this subject

Example: Notebook, Android, SSD hard drive

Display all 176 comments.
This thread is closed for comments
Top Comments
  • 36 Hide
    ubercake , December 12, 2013 7:22 AM
    I'm on page 4, and I can't even contain myself.

    Tearing and input lag at 60Hz on a 2560x1440 or 2560x1600 has been the only reason I won't game on one. G-sync will get me there.

    This is awesome, outside-of-the-box thinking tech.

    I do think Nvidia is making a huge mistake by keeping this to themselves though. This should be a technology implemented with every panel sold and become part of an industry standard for HDTVs, monitors or other viewing solutions! Why not get a licensing payment for all monitors sold with this tech? Or all video cards implementing this tech? It just makes sense.
  • 23 Hide
    expl0itfinder , December 12, 2013 7:11 AM
    Competition, competition. Anybody who is flaming over who is better: AMD or nVidia, is clearly missing the point. With nVidia's G-Sync, and AMD's Mantle, we have, for the first time in a while, real market competition in the GPU space. What does that mean for consumers? Lower prices, better products.
  • 17 Hide
    gamerk316 , December 12, 2013 6:34 AM
    I consider Gsync to be the most important gaming innovation since DX7. It's going to be one of those "How the HELL did we live without this before?" technologies.
Other Comments
  • 17 Hide
    gamerk316 , December 12, 2013 6:34 AM
    I consider Gsync to be the most important gaming innovation since DX7. It's going to be one of those "How the HELL did we live without this before?" technologies.
  • 16 Hide
    wurkfur , December 12, 2013 7:01 AM
    I personally have a setup that handles 60+ fps in most games and just leave V-Sync on. For me 60 fps is perfectly acceptable and even when I went to my friends house where he had a 120hz monitor with SLI, I couldn't hardly see much difference.

    I applaud the advancement, but I have a perfectly functional 26 inch monitor and don't want to have to buy another one AND a compatible GPU just to stop tearing.

    At that point I'm looking at $400 to $600 for a relatively paltry gain. If it comes standard on every monitor, I'll reconsider.
  • 23 Hide
    expl0itfinder , December 12, 2013 7:11 AM
    Competition, competition. Anybody who is flaming over who is better: AMD or nVidia, is clearly missing the point. With nVidia's G-Sync, and AMD's Mantle, we have, for the first time in a while, real market competition in the GPU space. What does that mean for consumers? Lower prices, better products.
  • 7 Hide
    Anonymous , December 12, 2013 7:13 AM
    This needs to be not so proprietary for it to become a game changer. As it is, requiring a specific GPU and specific monitor with an additional price premium just isn't compelling and won't reach a wide demographic.

    Is it great for those who already happen to fall within the requirements? Sure, but unless Nvidia opens this up or competitors make similar solutions, I feel like this is doomed to be as niche as lightboost, Physx, and, I suspect, Mantle.
  • -6 Hide
    Jgriff , December 12, 2013 7:21 AM
    g sync tv pleeeeeeeeeease
  • 36 Hide
    ubercake , December 12, 2013 7:22 AM
    I'm on page 4, and I can't even contain myself.

    Tearing and input lag at 60Hz on a 2560x1440 or 2560x1600 has been the only reason I won't game on one. G-sync will get me there.

    This is awesome, outside-of-the-box thinking tech.

    I do think Nvidia is making a huge mistake by keeping this to themselves though. This should be a technology implemented with every panel sold and become part of an industry standard for HDTVs, monitors or other viewing solutions! Why not get a licensing payment for all monitors sold with this tech? Or all video cards implementing this tech? It just makes sense.
  • 2 Hide
    rickard , December 12, 2013 7:25 AM
    Could the Skyrim stuttering at 60hz w/ Gsync be because the engine operates internally at 64hz? All those Bethesda tech games drop 4 frames every second when vsync'd to 60hz which cause that severe microstutter you see on nearby floors and walls when moving and strafing. Same thing happened in Oblivion, Fallout 3, and New Vegas on PC. You had to use stutter removal mods in conjunction with the script extenders to actually force the game to operate at 60hz and smooth it out with vsync on.

    You mention it being smooth when set to 144hz with Gsync, is there any way you cap the display at 64hz and try it with Gsync alone (iPresentinterval=0) and see what happens then? Just wondering if the game is at fault here and if that specific issue is still there in their latest version of the engine.

    Alternatively I suppose you could load up Fallout 3 or NV instead and see if the Gsync results match Skyrim.
  • 5 Hide
    Old_Fogie_Late_Bloomer , December 12, 2013 7:25 AM
    I would be excited for this if it werent for Oculus Rift. I don't mean to be dismissive, this looks awesome...but it isn't Oculus Rift.
  • -2 Hide
    hysteria357 , December 12, 2013 7:28 AM
    Am I the only one who has never experienced screen tearing? Most of my games run past my refresh rate too....
  • 1 Hide
    Gundamfan , December 12, 2013 7:45 AM
    i play alot of mmo and tearing apear all the time . this is nice tech i cant wait for 2560x1440. whit G-Sync
  • -5 Hide
    jasonpwns , December 12, 2013 7:46 AM
    TLM: Physx is not that way anymore. The newest gen of consoles can take advantage of Physx now and I'm sure will. That's hardly niche.
  • 9 Hide
    ohim , December 12, 2013 7:54 AM
    I think this is a game changer only on setups that can`t reach 60 fps, being proprietary only to nvidia is kinda shot in the foot, just like phys-x there are some people that care about it but most don`t even ecknowledge it, technically it can run on anything, practically only on nvidia and being a monitor side gimmick i would see this as a monitor company tech thing not graphics card maker closed stuff. I`m more interested in Mantle than this, since it preaches about better multicore cpu performance and better fps on your hardware.

    Mantle (if it will be what they say ) - better CPU performance, better GPU performance, at some point Open Source!?!? , no need for a new monitor.

    G-Sync good on old hardware that can`t reach 60 fps, bad since you need a new monitor, so guys who can`t afford a better GPU will have to get a new monitor ?!?!?!
  • -1 Hide
    pricetag_geek , December 12, 2013 7:55 AM
    You guys should've tried Dead Island with G-Sync, the tearing in this game was really horrible.
  • -1 Hide
    Gundamfan , December 12, 2013 7:55 AM
    I cant wait for G sync on 2560x1440
  • 10 Hide
    Someone Somewhere , December 12, 2013 7:59 AM
    Needs to be open. If it's not in a standard, it won't take off. How many people want to rip open a display (likely voiding warranty) to install a pricey addon card, and then be strongly limited to what other equipment they can use it with.

    Get it standardised and into the DVI/HDMI/DP specs, then it'll take off.

    I wonder if you could just add a flag for variable vertical blanks, and have it send a 'starting next frame' sequence whenever a frame is rendered.

    If it's not included by default in monitors, it'll become the next PhysX. And to do that it has to be platform-agnostic.
  • 2 Hide
    pricetag_geek , December 12, 2013 8:01 AM
    Quote:
    I personally have a setup that handles 60+ fps in most games and just leave V-Sync on. For me 60 fps is perfectly acceptable and even when I went to my friends house where he had a 120hz monitor with SLI, I couldn't hardly see much difference.


    oh really? I envy your eyes.
  • 2 Hide
    ubercake , December 12, 2013 8:09 AM
    Quote:
    I think this is a game changer only on setups that can`t reach 60 fps, being proprietary only to nvidia is kinda shot in the foot, just like phys-x there are some people that care about it but most don`t even ecknowledge it, technically it can run on anything, practically only on nvidia and being a monitor side gimmick i would see this as a monitor company tech thing not graphics card maker closed stuff. I`m more interested in Mantle than this, since it preaches about better multicore cpu performance and better fps on your hardware.

    Mantle (if it will be what they say ) - better CPU performance, better GPU performance, at some point Open Source!?!? , no need for a new monitor.

    G-Sync good on old hardware that can`t reach 60 fps, bad since you need a new monitor, so guys who can`t afford a better GPU will have to get a new monitor ?!?!?!


    Considering mantle, what does GPU performance matter on a screen with input lag or screen with tearing, choppy and blurry video?

    Mantle will not solve this problem. Mantle is supposed to be more of a low-level common API with enhanced GPU performance as a possible advantage. I'm not sure that even compares to what's being discussed here. Maybe I'm way off???

    G-sync will eliminate input lag, tearing and blur and as a result add to the overall realism of the gaming experience.
  • 2 Hide
    Logsdonb , December 12, 2013 8:27 AM
    This will become more important as we migrate to 4k displays. At that resolution, maintaining very high frame rates will become more difficult. Allowing a better experience at lower frame rates will become more important and more valuable.
  • -1 Hide
    renz496 , December 12, 2013 8:47 AM
    Quote:
    I think this is a game changer only on setups that can`t reach 60 fps, being proprietary only to nvidia is kinda shot in the foot, just like phys-x there are some people that care about it but most don`t even ecknowledge it, technically it can run on anything, practically only on nvidia and being a monitor side gimmick i would see this as a monitor company tech thing not graphics card maker closed stuff. I`m more interested in Mantle than this, since it preaches about better multicore cpu performance and better fps on your hardware.

    Mantle (if it will be what they say ) - better CPU performance, better GPU performance, at some point Open Source!?!? , no need for a new monitor.

    G-Sync good on old hardware that can`t reach 60 fps, bad since you need a new monitor, so guys who can`t afford a better GPU will have to get a new monitor ?!?!?!


    the monitor might be expensive right now but it will be good investment if you decide to go that route. at the very least you don't upgrade your monitor as often as gpu. my current monitor has been paired with GTS250, GTX460 and now GTX660 SLI. the only downside is it will locked you to use nvidia gpu only.
Display more comments