The Announcement
At Hot Chips 2025, NVIDIA introduced Spectrum-XGS Ethernet, a new “scale-across” networking technology designed to connect distributed data centers into what the company calls giga-scale AI super-factories. Unlike traditional scale-up or scale-out approaches that optimize within a single facility, Spectrum-XGS aims to merge multiple geographically dispersed data centers into a unified, high-performance cluster.
The technology integrates with NVIDIA’s existing Spectrum-X platform, introducing algorithms for adaptive congestion control, latency management, and telemetry tuned for long-distance connectivity. NVIDIA claims the system can nearly double the performance of its Collective Communications Library (NCCL), improving multi-GPU and multi-node communication even across facilities separated by cities or countries.
CoreWeave, a hyperscale AI infrastructure provider, will be one of the first to deploy the new technology, linking its distributed data centers into a single operational environment. Read the full press release here for more information.
Why the Industry Needs “Scale-Across”
As AI workloads grow more compute-intensive, enterprises and hyperscalers face physical and operational ceilings inside individual facilities. Power delivery, cooling capacity, and floor space are increasingly limiting how much infrastructure can be packed into one data center.
According to our research, more than 70% of organizations cite energy consumption and facility limits as barriers to scaling AI projects. At the same time, AI model sizes and dataset requirements are expanding at a pace that makes isolated clusters less viable.
Traditional Ethernet was never built for the ultra-low latency and predictable performance that AI workloads demand. Standard switching fabrics introduce jitter and congestion that can cripple multi-node training jobs or distributed inference pipelines. By extending Spectrum-X to interconnect facilities, NVIDIA is proposing a new dimension of scalability:
- Scale-up: Make individual nodes more powerful (larger GPUs, more memory).
- Scale-out: Add more nodes within a single facility.
- Scale-across: Interconnect facilities into one logical AI cluster.
This “third pillar” has significant implications for how future AI infrastructure will be architected.
Developer Implications of Distributed AI Fabric
For developers, Spectrum-XGS could mark a shift in how AI applications can be deployed and scaled.
- Predictable Training Across Regions: Distributed model training can now extend across multiple sites with less performance penalty, reducing the need to colocate all GPUs in one facility.
- Unified Resource Pools: Developers may treat geographically separated data centers as a single AI fabric, improving resource allocation for large jobs.
- Lower Latency Pipelines: With congestion control tuned for distance, real-time or near-real-time AI applications (recommendation engines, fraud detection, agentic AI workflows) can span distributed infrastructures.
- Resilience and Redundancy: Applications gain higher resilience by distributing workloads across multiple facilities without sacrificing coordination or speed.
Analysts at theCUBE Research have noted that as enterprises transition to AI-native application development, network performance is increasingly a gating factor. Scale-across infrastructure helps remove that barrier, enabling developers to focus on building agentic workflows, RAG pipelines, and other advanced systems without being constrained by physical site limits.
The CoreWeave Example of Hyperscale in Action
CoreWeave’s decision to adopt Spectrum-XGS highlights how hyperscalers view AI infrastructure differently from traditional enterprises. Instead of building the “largest single data center,” they are interconnecting multiple mid-sized facilities into a single logical supercomputer.
This approach matters for developers because it could offer:
- Elasticity across sites: Developers can request GPU resources without needing to know which data center provides them.
- Consistency: The network fabric hides geographic distance, reducing the variability in training and inference runs.
- Access to Giga-Scale Compute: Startups and enterprises alike can tap into infrastructure that feels like one colossal cluster, even though it is distributed across cities or continents.
Industry Impact of Changing AI Factories
The rise of AI factories, hyperscale facilities purpose-built for model training and inference, is reshaping data center economics. Our research points out that AI compute intensity is already straining traditional IT infrastructure, with electricity consumption projected to grow 4% annually through 2027, the fastest pace in recent history.
Spectrum-XGS aims to address this challenge by:
- Allowing organizations to distribute power and cooling demands across multiple facilities rather than overloading one.
- Reducing operational costs tied to network inefficiencies and retry traffic.
- Supporting multi-tenant environments where enterprises can securely run workloads at predictable performance levels.
From an industry perspective, this also blurs the line between cloud regions and clusters. In practice, hyperscalers and enterprises adopting this model could run AI workloads across distributed infrastructures as if they were a single site, a fundamental rethinking of what a “data center” means in the AI era.
Why This Matters
For developers, the promise of Spectrum-XGS is reliable, distributed scale. Training larger models or orchestrating agentic AI workflows no longer requires a single mega-facility; instead, the infrastructure itself adapts to geographic distribution.
For enterprises, the technology offers a path forward as they confront the dual challenges of AI growth and facility constraints. Rather than abandoning on-premises or colocation strategies in favor of hyperscale public cloud, Spectrum-XGS could enable hybrid AI architectures that span owned facilities and partner data centers.
The industry is at an inflection point where networking fabric matters as much as compute power. With Spectrum-XGS, NVIDIA is pushing the conversation beyond GPUs and into the realm of distributed systems design, signaling that the future of AI won’t be confined to a single building; it will be built on connected, intelligent fabrics that unify resources at giga-scale.