TL;DR:

  • NVIDIA has unveiled NVIDIA® Spectrum-XGS Ethernet, a new “scale-across” technology that combines multiple, geographically distributed data centers into a single, unified AI super-factory.
  • This technology is a breakthrough addition to the NVIDIA Spectrum-Xâ„¢ Ethernet platform, designed to overcome the physical limitations of single data centers.
  • Spectrum-XGS Ethernet features advanced algorithms, distance congestion control, and precision latency management to nearly double the performance of multi-GPU and multi-node communication over long distances.
  • CoreWeave, a leading AI infrastructure provider, will be one of the first to deploy the technology, connecting its data centers to form a unified supercomputer.
  • The technology is available now as part of the NVIDIA Spectrum-X Ethernet platform.
image of NVIDIA Introduces Spectrum-XGS Ethernet to Link Distributed Data Centers into Giga-Scale AI Super-Factories - HelloExpress - 2

NVIDIA’s “Scale-Across” Technology Creates Unified, Giga-Scale AI Super-Factories

HOT CHIPS — In a major announcement today, NVIDIA introduced NVIDIA® Spectrum-XGS Ethernet, a groundbreaking technology that allows for the creation of “giga-scale AI super-factories” by seamlessly linking distributed data centers. This innovation addresses the growing challenge of power and capacity limits within a single data center, which can hinder the scaling of AI workloads.

image of NVIDIA Introduces Spectrum-XGS Ethernet to Link Distributed Data Centers into Giga-Scale AI Super-Factories - HelloExpress - 3

“The AI industrial revolution is here, and giant-scale AI factories are the essential infrastructure,” said Jensen Huang, founder and CEO of NVIDIA. “With NVIDIA Spectrum-XGS Ethernet, we add scale-across to scale-up and scale-out capabilities to link data centers across cities, nations and continents into vast, giga-scale AI super-factories.”

Overcoming Traditional Network Limitations

Traditional off-the-shelf Ethernet networking is often insufficient for interconnecting geographically distributed data centers due to high latency, jitter, and unpredictable performance. Spectrum-XGS Ethernet solves this problem by integrating advanced algorithms that dynamically adapt the network to the distance between facilities.

image of NVIDIA Introduces Spectrum-XGS Ethernet to Link Distributed Data Centers into Giga-Scale AI Super-Factories - HelloExpress - 3

The technology features auto-adjusted distance congestion control and precision latency management, which work to accelerate multi-GPU and multi-node communication. As a result, Spectrum-XGS Ethernet nearly doubles the performance of the NVIDIA Collective Communications Library, enabling multiple data centers to operate as a single, fully optimized AI super-factory.

CoreWeave to Be a First Adopter

Hyperscale pioneers are already embracing this new infrastructure. CoreWeave, a specialized AI cloud provider, will be among the first to deploy NVIDIA Spectrum-XGS Ethernet.

“With NVIDIA Spectrum-XGS, we can connect our data centers into a single, unified supercomputer, giving our customers access to giga-scale AI that will accelerate breakthroughs across every industry,” said Peter Salanki, cofounder and chief technology officer of CoreWeave.

Spectrum-XGS Ethernet is available now as part of the NVIDIA Spectrum-X Ethernet platform, which also includes NVIDIA Spectrum-X switches and NVIDIA ConnectX®-8 SuperNICs, all designed to deliver seamless scalability and ultralow latency for the most demanding AI workloads.

You may also like

Leave a reply

Your email address will not be published. Required fields are marked *