Micron Unveils 128GB and 256GB CXL 2.0 Expansion Modules

Micron
(Image credit: Micron)

Micron on Monday introduced its CZ120 memory expansion modules that comply with the CXL 2.0 Type 3 specification and feature a PCIe 5.0 x8 interface. The modules are designed to expand DRAM capacity and bandwidth for servers that need more high-performance memory to run workloads that require loads of RAM, including in-memory databases and software-as-a-service. Samples of the modules are now available to interested parties.

The Micron CZ120 memory modules are equipped with 128GB and 256GB of memory, which is in line with what typical RDIMMs offer. The expansion modules come in a E3.S 2T form-factor with a PCIe 5.0 x8 interface and are based on Microchip's SMC 2000 controller compliant with the CXL 2.0 Type 3 standard as well as Micron's memory chips made using the company's 1α (1-alpha) DRAM production node. 

From a performance point of view, Micron's CZ120 CXL 2.0 memory expansion modules provide bandwidth of up to 36 GB/s (measured by running MLC workload with 2:1 read/write ratio on a single CZ120 memory expansion module) and are not significantly slower than DDR5-4800 RDIMMs with 128GB and 256GB memory onboard that offer a peak bandwidth of 38.4 GB/s.  

"Micron is advancing the adoption of CXL memory with this CZ120 sampling milestone to key customers," said Siva Makineni, vice president of the Micron Advanced Memory Systems Group. "We have been developing and testing our CZ120 memory expansion modules utilizing both Intel and AMD platforms capable of supporting the CXL standard. Our product innovation coupled with our collaborative efforts with the CXL ecosystem will enable faster acceptance of this new standard, as we work collectively to meet the ever-growing demands of data centers and their memory-intensive workloads."

TOPICS
Anton Shilov
Contributing Writer

Anton Shilov is a contributing writer at Tom’s Hardware. Over the past couple of decades, he has covered everything from CPUs and GPUs to supercomputers and from modern process technologies and latest fab tools to high-tech industry trends.

  • bit_user
    I get the sense they're holding back, while they test the waters. I think you should probably be able to pack more capacity in that form factor, even before resorting to chip-stacking.

    BTW, I think both Genoa and Sapphire Rapids only go as high as CXL 1.1. I wonder if Emerald Rapids is going to move up to CXL 2.0. The difference isn't raw speed, but rather a matter of features and functionality. You have to go all the way to CXL 3.0, before there's a bump in speed (i.e. PCIe 6.0 PHY).
    Reply
  • thestryker
    bit_user said:
    I get the sense they're holding back, while they test the waters. I think you should probably be able to pack more capacity in that form factor, even before resorting to chip-stacking.
    I'm betting that this is likely the best they can do with 16Gb chips even using both sides of the PCB. I would assume that a move to 24Gb would happen before stacking, and maybe even 32Gb depending on if they hit manufacturing timelines.
    bit_user said:
    BTW, I think both Genoa and Sapphire Rapids only go as high as CXL 1.1. I wonder if Emerald Rapids is going to move up to CXL 2.0. The difference isn't raw speed, but rather a matter of features and functionality. You have to go all the way to CXL 3.0, before there's a bump in speed (i.e. PCIe 6.0 PHY).
    Intel has been pretty cagey about EMR specs so I'd assume it's 1.1 again while Granite/Sierra should be 2.0. With any luck there will be details at Hotchips, but if not I'd imagine there should be some during the Innovation event.
    Reply
  • bit_user
    thestryker said:
    Intel has been pretty cagey about EMR specs so I'd assume it's 1.1 again while Granite/Sierra should be 2.0.
    I think I might have seen a leaked roadmap, and it's as you say.

    IIRC, 2.0 adds switching (or maybe it's multi-level switching?), and that will enable much larger memory capacities by allowing more of these devices than there are CXL lanes to support them. 3.0 adds support for fabrics, and that should be an enabler for having shared RAM pools at rack-scale.
    Reply
  • Li Ken-un
    CXL would have been the perfect use for 3D XPoint (Optane) media:
    Denser than RAM, and thus higher-capacity modules;
    But also byte-addressable like RAM; and
    Both faster and lower latency than NANDToo bad we’re stuck with either the proprietary DIMMs which work only in special Intel proprietary systems or slower PCIe-limited SSDs.
    Reply
  • bit_user
    Li Ken-un said:
    CXL would have been the perfect use for 3D XPoint (Optane) media:
    Intel were supposedly working on such a product, when Optane got canceled.

    Li Ken-un said:

    Denser than RAM, and thus higher-capacity modules;
    But also byte-addressable like RAM; and
    Both faster and lower latency than NAND
    You can do almost as well (in some respects, better) with NAND-backed DRAM + power-loss capacitors. Optane was denser and cheaper than DRAM, but not much. It wouldn't offer a capacity advantage over die-stacked DDR5.

    The biggest problem with Optane DIMMs was extremely limited software support for using it as persistent memory. If you use it like a fast SSD, then you lose the benefits of byte-addressability and kernel overhead kills most of the performance advantage vs. just having an Optane NVMe SSD.
    Reply