China makes AI breakthrough, reportedly trains generative AI model across multiple data centers and GPU architectures
Necessity is the mother of invention.
An industry analyst recently revealed that China has developed a single generative AI (GAI) model across multiple data centers — a massive feat considering the complexity of using different GPUs in a single data center, let alone using servers in multiple geographic locations. Patrick Moorhead, Chief Analyst at Moor Insights & Strategy, said on X (formerly Twitter) that China was the first country to manage this achievement and that he discovered it during a conversation about a presumably unrelated NDA meeting.
This technique of training GAIs across different locations/architectures is essential for China to keep its AI dreams moving forward, especially as American sanctions have stopped it from acquiring the latest, most powerful chips to drive its research and development. Since Nvidia does not want to lose the Chinese market, it created the less powerful H20 AI chips that fall within Washington’s restrictive performance parameters. However, there are rumors that even these down-tuned chips might be banned soon, highlighting the uncertainty Chinese tech companies face in the current political climate.
Because of this uncertainty, Chinese researchers have been working on melding GPUs from different brands into one training cluster. By doing so, the institutions could combine their limited stocks of sanctioned high-end, high-performance chips, like the Nvidia A100, with less powerful but readily available GPUs, like Huawei’s Ascend 910B or the aforementioned Nvidia H20. This technique could help them combat the high-end GPU shortage within China, although it has historically come with large drops in efficiency.
However, it seems that China has found ways to solve this issue, especially with the news of the single GAI development across multiple data centers. Although we don’t have any information on this GAI yet, it shows the lengths that Chinese researchers will go to, to ensure that they can continue driving China’s AI ambitions forward. As Huawei said, China would find ways to continue moving its AI development despite American sanctions. After all, necessity is the mother of invention.
Stay On the Cutting Edge: Get the Tom's Hardware Newsletter
Get Tom's Hardware's best news and in-depth reviews, straight to your inbox.
Jowi Morales is a tech enthusiast with years of experience working in the industry. He’s been writing with several tech publications since 2021, where he’s been interested in tech hardware and consumer electronics.
-
Notton Is it a breakthrough though?Reply
AFAIK using multiple data centers is not a new thing, it's just that no one likes dealing with the latency delays of waiting for data to arrive over the internet.
Which is why the preferred method is to put all eggs in one basket. -
usertests
If their technique allows the training to be broken up into latency-insensitive chunks, like Folding@home, then it should probably be considered a breakthrough. I don't know if that's possible, just throwing it out there.Notton said:Is it a breakthrough though?
AFAIK using multiple data centers is not a new thing, it's just that no one likes dealing with the latency delays of waiting for data to arrive over the internet.
Which is why the preferred method is to put all eggs in one basket. -
zsydeepsky Notton said:Is it a breakthrough though?
AFAIK using multiple data centers is not a new thing, it's just that no one likes dealing with the latency delays of waiting for data to arrive over the internet.
Which is why the preferred method is to put all eggs in one basket.
well, I remember reading news from MS and FB, both mentioning they met a bottleneck in training AI within a single data center.
that bottleneck is electricity.
the requirement for electricity was so huge that a data center could consume gigawatts, the single data center workload alone can destabilize the power grid.
I would say that the US developers will eventually have no choice but to train AI in multiple data centers as well. -
acadia11
Or build dedicated data center power plants like Oracle new planned nuclear powered data center, or Microsoft’s repurposing of 3 mile island. The real breakthrough will be cold fusion … interesting thought experiment what would be the impact of limitless power supply on our economic and technical endeavors as a society?zsydeepsky said:well, I remember reading news from MS and FB, both mentioning they met a bottleneck in training AI within a single data center.
that bottleneck is electricity.
the requirement for electricity was so huge that a data center could consume gigawatts, the single data center workload alone can destabilize the power grid.
I would say that the US developers will eventually have no choice but to train AI in multiple data centers as well. -
usertests
Limitless? We would turn Earth into magma.acadia11 said:Or build dedicated data center power plants like Oracle new planned nuclear powered data center, or Microsoft’s repurposing of 3 mile island. The real breakthrough will be cold fusion … interesting thought experiment what would be the impact of limitless power supply on our economic and technical endeavors as a society?
There's still costs associated with it. But if we end up with $0.01/kWh, it will allow us to do some truly wasteful stuff. -
Li Ken-un
AI will be the impetus for building Dyson swarms around the sun. lolacadia11 said:The real breakthrough will be cold fusion … interesting thought experiment what would be the impact of limitless power supply on our economic and technical endeavors as a society? -
Gururu Not sure that overcoming a state-specific deficiency should be considered an industry-wide breakthrough unless it is a must-be adopted practice across said industry. Time will tell I suppose.Reply -
The Historical Fidelity Love these types of articles based on X posts of a guy who overheard someone say it happened, but it was in a meeting where he can’t talk about it due to an NDA, which means he overheard it being talked about during unrelated banter in the NDA meeting, which means there is 0 confirmability on the authenticity of this breakthrough other than hearsay…Reply -
Geef Li Ken-un said:AI will be the impetus for building Dyson swarms around the sun. lol
Here is a link for an Isaac Arthur video about Matrioshka Brains. -
TCA_ChinChin
Unfortunately this feels like 50% of all articles nowadays, and not just for tech related stuff.The Historical Fidelity said:Love these types of articles based on X posts of a guy who overheard someone say it happened, but it was in a meeting where he can’t talk about it due to an NDA, which means he overheard it being talked about during unrelated banter in the NDA meeting, which means there is 0 confirmability on the authenticity of this breakthrough other than hearsay…