- Articles & News
- For IT Pros
- Your Opinion
We tested at the maximum graphics quality and the highest illumination model. Extended dynamic illumination of objects (DirectX 10), anti-aliasing, and anisotropic filtering were deactivated and the range of view was set to maximum. The test system consists of a Core 2 Duo E6750 CPU at its standard frequency of 2.66 GHz, 4 GB of memory, and a GeForce 8800 GTS 512.
The question that keeps coming up is what is a monitor with widescreen resolution good for? If a game supports this resolution, there won’t be any distortion of width or black borders. The following pictures show the visible area and resolution, and the aspect ratio is next to it. You can see the achievable frame rate in the top left corner. All 16:10 formats have a wide image and show more of the picture. The 16:9 format has the widest image, which is a significant advantage because you see even more of the environment.
In the table you’ll see two percentages. The baseline of the first one is 800x600 pixels and it shows the loss in frame rate that occurs as you change to higher resolutions. The baseline of the second percentage is 1920x1200 pixels and it illustrates the gain in percentage terms when you decrease resolution.
The test scene we chose is challenging for the graphics card. The optimal setting for graphics quality would be 1280x1024 pixels. If you own a monitor with a resolution of 1920x1200 pixels, you will have to set the resolution to 1440x900 pixels to stay within your aspect ratio and still have a smooth frame rate for maximum graphics quality. You only gain 20 percent in performance when changing to 1680x1050. Given the baseline of 15 fps, this is not enough to exceed 25 fps.
|Resolution in Pixels||fps||Percent||Percent|