Microsoft Azure has unveiled the world’s first production-scale NVIDIA GB300 NVL72 supercomputing cluster, specifically designed for OpenAI’s most demanding AI workloads. This milestone represents a significant leap in AI infrastructure capabilities, featuring over 4,600 NVIDIA Blackwell Ultra GPUs that will power next-generation reasoning models and agentic AI systems while reinforcing American leadership in artificial intelligence.
What you should know: The new NDv6 GB300 VM series delivers unprecedented computational power through liquid-cooled, rack-scale systems that integrate cutting-edge hardware into a unified supercomputer.
- Each rack contains 72 NVIDIA Blackwell Ultra GPUs and 36 NVIDIA Grace CPUs, providing 37 terabytes of fast memory and 1.44 exaflops of FP4 Tensor Core performance per VM.
- The cluster connects over 4,600 GPUs using NVIDIA’s Quantum-X800 InfiniBand networking platform, creating a massive unified memory space essential for complex AI model training and inference.
- Microsoft applied “radical engineering to memory and networking” to achieve the massive scale of compute required for high-performance reasoning models.
Performance breakthrough: Recent benchmarks demonstrate the system’s exceptional capabilities, particularly for large language models and reasoning AI.
- In MLPerf Inference v5.1 benchmarks, the GB300 NVL72 systems delivered up to 5x higher throughput per GPU on the 671-billion-parameter DeepSeek-R1 reasoning model compared to NVIDIA’s previous Hopper architecture.
- The platform achieved leadership performance across all newly introduced benchmarks, including the Llama 3.1 405B model.
- NVIDIA Blackwell Ultra supports new formats like NVFP4 for breakthrough training performance and compiler technologies like NVIDIA Dynamo for optimal inference performance.
The networking architecture: A sophisticated two-tiered approach enables both scale-up performance within racks and scale-out performance across the entire cluster.
- Within each GB300 NVL72 rack, fifth-generation NVIDIA NVLink Switch fabric provides 130 TB/s of direct, all-to-all bandwidth between the 72 Blackwell Ultra GPUs.
- The NVIDIA Quantum-X800 InfiniBand platform delivers 800 Gb/s of bandwidth per GPU, featuring ConnectX-8 SuperNICs and advanced adaptive routing capabilities.
- NVIDIA Scalable Hierarchical Aggregation and Reduction Protocol (SHARP) v4 accelerates operations to significantly boost efficiency of large-scale training and inference.
Why this matters: This deployment marks a crucial step toward building the infrastructure needed for frontier AI development and positions the United States at the forefront of AI innovation.
- The achievement required reimagining every layer of Microsoft’s data center infrastructure, from custom liquid cooling and power distribution to reengineered software stacks.
- As Azure scales toward deploying hundreds of thousands of NVIDIA Blackwell Ultra GPUs, this milestone enables customers like OpenAI to access unprecedented computational resources for developing next-generation AI systems.
What they’re saying: Industry leaders emphasize the collaborative engineering effort required to achieve this scale of AI infrastructure.
- “Delivering the industry’s first at-scale NVIDIA GB300 NVL72 production cluster for frontier AI is an achievement that goes beyond powerful silicon — it reflects Microsoft Azure and NVIDIA’s shared commitment to optimize all parts of the modern AI data center,” said Nidhi Chappell, corporate vice president of Microsoft Azure AI Infrastructure.
- “Our collaboration helps ensure customers like OpenAI can deploy next-generation infrastructure at unprecedented scale and speed.”
Recent Stories
DOE fusion roadmap targets 2030s commercial deployment as AI drives $9B investment
The Department of Energy has released a new roadmap targeting commercial-scale fusion power deployment by the mid-2030s, though the plan lacks specific funding commitments and relies on scientific breakthroughs that have eluded researchers for decades. The strategy emphasizes public-private partnerships and positions AI as both a research tool and motivation for developing fusion energy to meet data centers' growing electricity demands. The big picture: The DOE's roadmap aims to "deliver the public infrastructure that supports the fusion private sector scale up in the 2030s," but acknowledges it cannot commit to specific funding levels and remains subject to Congressional appropriations. Why...
Oct 17, 2025Tying it all together: Credo’s purple cables power the $4B AI data center boom
Credo, a Silicon Valley semiconductor company specializing in data center cables and chips, has seen its stock price more than double this year to $143.61, following a 245% surge in 2024. The company's signature purple cables, which cost between $300-$500 each, have become essential infrastructure for AI data centers, positioning Credo to capitalize on the trillion-dollar AI infrastructure expansion as hyperscalers like Amazon, Microsoft, and Elon Musk's xAI rapidly build out massive computing facilities. What you should know: Credo's active electrical cables (AECs) are becoming indispensable for connecting the massive GPU clusters required for AI training and inference. The company...
Oct 17, 2025Vatican launches Latin American AI network for human development
The Vatican hosted a two-day conference bringing together 50 global experts to explore how artificial intelligence can advance peace, social justice, and human development. The event launched the Latin American AI Network for Integral Human Development and established principles for ethical AI governance that prioritize human dignity over technological advancement. What you should know: The Pontifical Academy of Social Sciences, the Vatican's research body for social issues, organized the "Digital Rerum Novarum" conference on October 16-17, combining academic research with practical AI applications. Participants included leading experts from MIT, Microsoft, Columbia University, the UN, and major European institutions. The conference...