Important Considerations For 3D Video
Should I upgrade my TV, or my PC?
Ultimately, you’ll want to watch movies on the largest screen that you can afford. Several 3D TV models are available today, and more will be available later this year.
This year, 3D TVs are going to be relatively expensive. Typically, the replacement cycle for TVs is between five and 10 years. Consumers who have recently purchased a new large-screen TV may be reluctant to upgrade to a new 3D TV right away. It is likely that consumers will add 3D capability at some point in the future when they otherwise choose to upgrade or replace their TV. Of course, this decision depends on many factors, such as the availability of Blu-ray 3D titles, 3D TV channels, and other 3D video content.
Replacement cycles for notebook PCs, and upgrade cycles for desktop PCs are much faster. Enthusiasts may upgrade their desktop PCs every year. It will be easy to add 3D video decoding and display capability when upgrading or replacing a PC. For these reasons, we think that the installed base of 3D video-capable PCs will vastly outnumber 3D capable TVs in the next few years.
Decoding Blu-ray 3D on a PC
While quad-core CPUs can support software decoding of 3D Blu-ray, the optimal solution includes a discrete graphics card or integrated graphics solution capable of decoding Blu-ray 3D in the GPU. The latest-generation graphics processors, including Nvidia’s GeForce GT 240, GT 340, GT 330, GT 320, GTX 470, GTX 480 graphics cards, and GeForce 300M-series mobile graphics, and systems with Intel Core processors with Intel HD graphics (Core i3, Core i5, and Core i7 Mobile) support dual HD video stream decoding. Blu-ray 3D video decoding solutions can be expected for ATI Radeon 5000-series graphics in the future.
Blu-ray player software utilizes these modern graphics processors to decode Blu-ray 3D MVC, resulting in very low CPU utilization and flawless video performance.
Connecting To A PC display
Full-quality 120 Hz frame sequential 3D video is only supported through a dual-link DVI connector (for Nvidia 3D Vision-compatible displays), or (soon) through a High Speed HDMI cable to a HDMI 1.4-compliant display.
HDMI 1.4 specifies support for a number of 3D video signal formats, including full-frame, dual-stream 3D, where both the left and right video frames are packed into a single stereo frame, with the left eye picture on top of the right. HDMI 1.4 stereoscopic frame packing supports 1080p at 24 frames per second, or 720p resolution at 50 or 60 frames per second.
HDMI 1.4 also defines 3D signals compressed into standard 2D video formats, including side-by-side and over/under. Polarized displays can be connected to a PC using standard DVI or HDMI 1.3 connections.
Unfortunately I am one of those who recently purchased a new TV. It will be quite some time before I upgrade.
I understand the author's preference for shutter glasses (especially since it's a certain product's preferred method of choice) even if I don't share it, the major limitation is having to buy a pair for all your friends coming over, which gets impractical until they are more commonplace.
Also polarized solutions are not limited in resolution if they are set-up beyond just the example provided in this article (like they do in the theatre with dual projectors ) and may have an improving single source future with 2K and 4K displays on the horizon. It's a question of preference, but it seems like the full story wasn't explored on that subject.
Now on to a pet peeve: I love the part about "While set-top Blu-ray players will need to be replaced, PC-based Blu-ray player software can be upgraded." as a subtle product benefit plug.
Unless it's a free upgrade, you are still replacing the software, not upgrading it (it's not a plug-in), and you're likely forking out nearly the same amount of money for the 1/100th of the cost to produce that software update, so it's not like it's a major advantage. Especially when upgrading requires a FULL upgrade to the most expensive model Power DVD (version #) Ultra 3D, and I can't simply add it to my existing PowerDVD bundles thus potentially changing my backwards compatibility (Ultra 9 already removed my HD-DVD support from Ultra 7 that I upgraded on my LG HD-DVD/BR burner
, until then it's $99 (or $94.95 for loyal saps) vs $150-200, plus with the set-top route now I have a second BR-/DVD player for another room or to give to a friend (the BR software on its own is useless to give to someone else without a drive), and that's not even compared to the free PS3 upgrade.
Also can someone explain this statement;
"Blu-ray 3D video decoding solutions can be expected for ATI Radeon 5000-series graphics in the future."
Didn't Cyberlink already show their BR-3D solution on ATi hardware last year? So what's the issue?
Also why is it limited to "GeForce 300M-series mobile graphics" when often the core is the same a previous generation 200M series (example GTS 350M / 250M )?
And this section "Full-quality 120 Hz frame-sequential 3D video (such as Blu-ray 3D) is only supported through a High Speed
HDMI cable to a HDMI 1.4-compliant TV. " seems to miss the DVI dual-link to monitor option currently being used for 3D on PCs, and also the dual 1.3 input monitors/TVs.
A nice little article for people unfamiliar with 3D, but there's a subtle under-current of product preference/placement in it, and far too many generalities with little supporting information. :??:
also some info on AVRs and how a 1.3 hdmi AVR might pass on 3d video and still decode bitstream audio, or not - do we need 1.4 hdmi AVRs to decode audio from a 1.4 source? we shouldn't need 1.4 receivers since the audio standards haven't change, but I'm understanding that in fact we do neeed new receivers. :/
This piece is a prelude to tomorrow's coverage, by Don, of Blu-ray 3D on a notebook and a desktop. Perhaps that one will answer any of the questions you were left with here?
As for AMD, Tom and I went back and forth on this piece, and we agreed that it was critical to get AMD's feedback on Blu-ray 3D readiness. The fact of the matter is that it isn't ready to discuss the technology. It's behind.
The mention of dual-link DVI was in the first revision of this piece and removed in a subsequent iteration. I've asked the author for additional clarification there and should have an answer shortly.
It turns out the demo (I think it was at CES?) only used CPU decoding over an ATI graphics card; the Radeon did no software decoding.
The Cyberlink rep tells me that Blu-ray 3D software decoding is extremely CPU-dependant and might even require a quad-core CPU. He said all four threads were being stressed under software decoding, not sure what quad-core CPU they were using though.
Definitely something I'd like to test out in the future...
These implementations, while ever more impressive, are still being threshed out. Because of possible physiological side effects, I think I will NOT be a first adopter, with this (particular) tech (3D).
Anyone ever watch that movie "THE JERK", with STEVE MARTIN ??
= Opti-Grab =
... I can see all these class-action suits by parents of cross-eyed gamers ... hope not, tho ... I *AM* very much looking forward to the fully refined "end game", for 3D ...
Additionally, the very best desktop workstations are only just now catching up to standard (uncompressed) HD resolution ingest and edit/render ... since that bandwidth IS shared, between both eyes, this may be a non-issue.
I will let the kiddies and 1st adopters take-on all those risks and costs.
Please let me know when it is all "fully baked" and field tested!
= Alvin = (not to mention "affordable").