Nvidia's GeForce RTX 2050, MX570, MX550 GPUs Arrive In Spring 2022

GeForce MX570
GeForce MX570 (Image credit: Nvidia)

Update 12/17/2021 10:19 PT:

Nvidia has confirmed to German news outlet Computerbase the architectures behind the GeForce RTX 2050, MX570 and MX550 graphics cards. We've amended the story to reflect the changes.

Original Article

Nvidia slipped out a quiet announcement of the arrival of three new mobile graphics cards in the spring of 2022. The GeForce RTX 2050 will join the ranks of the GeForce MX570 and MX550 in powering a new wave of laptops for gamers and creators.

The GeForce RTX 2050 and GeForce MX570 are based on the GA107 (Ampere) silicon, the same silicon that powers the GeForce RTX 3050 and RTX 3050 Ti Mobile. The GeForce MX550, on the other hand, is still based on the TU117 (Turing) silicon from the current GeForce MX450.

In regards to featureset, the GeForce MX570 will support Nvidia DLSS and ray tracing in a "limited" fashion. Although the GeForce MX550 has the Tensor and RT cores bake in it, the Turing-based graphics card do not support Nvidia DLSS or ray tracing. As a result, it doesn't support Nvidia Broadcast unlike the GeForce RTX 2050 and MX570. However, the trio of upcoming GeForce graphics card are compatible with the Resizable BAR and Optimus technologies, but not Advanced Optimus.

GeForce RTX 2050 Specifications

Swipe to scroll horizontally
Header Cell - Column 0 GeForce RTX 2080 SuperGeForce RTX 2080GeForce RTX 2070 SuperGeForce RTX 2070GeForce RTX 2060GeForce RTX 2050
Nvidia CUDA Cores3,0722,9442,5602,3041,9202,048
Boost Clock (MHz)1,080 - 1,560 MHz1,095 - 1,590 MHz1,155 - 1,380 MHz1,125 - 1,455 MHz1,185 - 1,560 MHz1,155 - 1,477 MHz
Graphics Subsystem Power (W)80 - 150+ W80 - 150+ W80 - 115 W80 - 115 W65 - 115 W30 - 45 W
Memory SpeedUp to 14 GbpsUp to 14 GbpsUp to 14 GbpsUp to 14 GbpsUp to 14 GbpsUp to 14 Gbps
Standard Memory Configuration8 GB GDDR68 GB GDDR68 GB GDDR68 GB GDDR66 GB GDDR64 GB GDDR6
Memory Interface Width256-bit256-bit256-bit256-bit192-bit64-bit
Memory Bandwidth (GB/sec)Up to 448 GB/sUp to 448 GB/sUp to 448 GB/sUp to 448 GB/sUp to 336 GB/sUp to 112 GB/s

Nvidia listed the GeForce RTX 2050 with 2,048 CUDA cores, around 7% more than the GeForce RTX 2060. However, the GeForce RTX 2050 adheres to a more restricted TDP so it'll come with lower clock speeds. The GeForce RTX 2060 should still be faster. As usual, Nvidia gives its partners some thermal headroom, meaning the power limit on the GeForce RTx 2050 will vary from one product to another. In any event, the graphics card is rated between 30W to 45W. Therefore, the boost clock will vary as well, and we'll find the GeForce RTX 2050 boosting between 1,155 MHz and 1,477 MHz.

In terms of the memory configuration, the GeForce RTX 2050 has 4GB of GDDR6 memory clocked at 14 Gbps. Bound by a 64-bit memory interface, the GeForce RTX 2050 can deliver a maximum memory bandwidth up to 112 GBps.

Meanwhile, the GeForce MX550 will be the direct replacement for the GeForce MX450, whereas the GeForce MX570 appears to be a new SKU. Nvidia didn't delve into the intricacies of the GeForce MX570 and MX550. The product pages confirm support for PCIe 4.0 and GDDR6 memory. 

The GeForce MX570 and MX550's other specifications remain a mystery. The chipmaker only stated that the duo sport "more power-efficient CUDA cores and with faster memory speeds than previous MX GPUs." In addition, Nvidia claims that the GeForce MX570 and MX550 will offer a leap of performance and efficiency in photo and video-editing tasks, such as Adobe Lightroom and Premiere Pro, as well as improved gaming in comparison to integrated graphics.

Zhiye Liu
News Editor and Memory Reviewer

Zhiye Liu is a news editor and memory reviewer at Tom’s Hardware. Although he loves everything that’s hardware, he has a soft spot for CPUs, GPUs, and RAM.

  • logainofhades
    Nvidia flooding the market with more products, to confuse the customer with.
    Reply
  • -Fran-
    logainofhades said:
    Nvidia flooding the market with more products, to confuse the customer with.
    Well, I will say its better than just rebranding them and making the audience think they're Ampere versions like they've done in the past (9800GT cough cough). Same for AMD, mind you.

    More than doing this, I personally hate shameless rebranding, so this is not too bad IMO. Plus, there is a need for more GPUs in all market segments, so they're just filling gaps I'd say.

    This does confuse things for some, but it's clear they're not Ampere cards, so that's good.

    Regards.
    Reply
  • hotaru.hino
    logainofhades said:
    Nvidia flooding the market with more products, to confuse the customer with.
    The average customer likely doesn't care enough to be able to differentiate a "RTX 2050" from a "RX 6600" anyway.
    Reply
  • hasten
    -Fran- said:
    Well, I will say its better than just rebranding them and making the audience think they're Ampere versions like they've done in the past (9800GT cough cough). Same for AMD, mind you.

    More than doing this, I personally hate shameless rebranding, so this is not too bad IMO. Plus, there is a need for more GPUs in all market segments, so they're just filling gaps I'd say.

    This does confuse things for some, but it's clear they're not Ampere cards, so that's good.

    Regards.
    I would argue this is a little different than nvidia and especially AMD/ATI using less than honest naming convention, but agree with your point. Although this time around they could have named it the 256 v2 and it would fly off the shelves...
    Reply
  • jacob249358
    Admin said:
    Nvidia expands its mobile offerings with the new GeForce RTX 2050, MX570, MX550 and graphics cards.

    Nvidia's GeForce RTX 2050, MX570, MX550 GPUs Arrive In Spring 2022 : Read more
    Why, why, why. Literally, no one wants anything other than 30 series on the medium/high end and 16 series on the low/budget end.
    Reply
  • logainofhades
    hotaru.hino said:
    The average customer likely doesn't care enough to be able to differentiate a "RTX 2050" from a "RX 6600" anyway.

    You would be surprised what people get confused over.
    Reply
  • linkup901
    "The GeForce RTX 2050 will...powering a new wave of laptops for gamers and creators."

    Those poor poor POOR gamers. I know the prices are insane, but this card is not it.
    Reply
  • lazyabum
    logainofhades said:
    Nvidia flooding the market with more products, to confuse the customer with.
    Oh yeah, like the shelves in stores everywhere are overflowing with GPUs.
    Reply
  • hotaru.hino
    logainofhades said:
    You would be surprised what people get confused over.
    Sure, but I would wager a lot of them would go find a sales rep if they don't know what to look for. And then said rep will just steer them to a particular model.

    If they're configuring a laptop on say Dell or whatever, the situation is usually:
    There's a configuration that's iGPU only or not
    The choice of dGPUs tends to be limited to a specific configuration of a model. All the person has to choose is which laptop fits their budget
    And if they run into a situation where multiple configurations are at the sameish price point, they'll probably choose what has the bigger numbers or what's familiar to them anyway. Or just wing it and be on their merry way because whatever they're getting is going to be an upgrade anyway.I don't think the average person cares about the hardware as long as it's feels better than what they came from and they don't perceive the manufacturer as rubbish. And if anyone does actually care, they probably can figure their way around this model number soup anyway.
    Reply
  • InvalidError
    jacob249358 said:
    Why, why, why. Literally, no one wants anything other than 30 series on the medium/high end and 16 series on the low/budget end.
    Nvidia doesn't want to support retired SKUs forever, got to push the market forward so the old stuff can be phased out from the market and drivers. Also, the low-end already had the RTX3050(Ti) on the mobile side, so the RTX2050 and below are most likely because Nvidia cannot make enough GA107s.
    Reply