Microsoft Deploys First Massive AI System, Signaling Expansion of Azure AI Capabilities
Microsoft yesterday announced the deployment of its first large-scale AI system – dubbed an “AI factory” – across its global Azure data center network, a move designed to bolster its capacity to support rapidly growing artificial intelligence workloads.
The newly deployed system comprises a cluster of over 4,600 Nvidia GB300 rack computers, each powered by the Nvidia Blackwell Ultra GPU chip and interconnected using Nvidia’s InfiniBand networking technology. Microsoft plans to deploy “hundreds of thousands” of these Blackwell Ultra GPUs as it expands these AI systems globally, positioning itself to handle next-generation AI models with “hundreds of trillions of parameters.” This investment is crucial as demand for AI processing power continues to surge.
The announcement follows recent data center deals between OpenAI – a key Microsoft partner – and both Nvidia and AMD, with OpenAI committing to an estimated $1 trillion in data center construction by 2025. Microsoft emphasized that its existing infrastructure, with over 300 data centers in 34 countries, is “uniquely positioned” to meet the current and future demands of advanced AI development. You can learn more about the growing field of AI data centers on Nvidia’s website.
Further details regarding Microsoft’s AI infrastructure expansion are expected later this month, with Chief Technology Officer Kevin Scott scheduled to speak at TechCrunch Disrupt in San Francisco from October 27-29. Microsoft’s Azure platform is a major competitor in the cloud computing market.
Microsoft CEO Satya Nadella on Thursday tweeted a video of his company’s first deployed massive AI system — or AI “factory” as Nvidia likes to call them. He promised this is the “first of many” such Nvidia AI factories that will be deployed across Microsoft Azure’s global data centers to run OpenAI workloads.
Each system is a cluster of more than 4,600 Nvidia GB300 rack computers sporting the much-in-demand Blackwell Ultra GPU chip and connected via Nvidia’s super-fast networking tech called InfiniBand. (Besides AI chips, Nvidia CEO Jensen Huang also had the foresight to corner the market on InfiniBand when his company acquired Mellanox for $6.9 billion in 2019.)
Microsoft promises that it will be deploying “hundreds of thousands of Blackwell Ultra GPUs” as it rolls out these systems globally. While the size of these systems is eye-popping (and the company shared plenty more technical details for hardware enthusiasts to peruse), the timing of this announcement is also noteworthy.
It comes just after OpenAI, its partner and well-documented frenemy, inked two high-profile data center deals with Nvidia and AMD. In 2025, OpenAI has racked up, by some estimates, $1 trillion in commitments to build its own data centers. And CEO Sam Altman said this week that more were coming.
Microsoft clearly wants the world to know that it already has the data centers — more than 300 in 34 countries — and that they are “uniquely positioned” to “meet the demands of frontier AI today,” the company said. These monster AI systems are also capable of running the next generation of models with “hundreds of trillions of parameters,” it said.
We expect to hear more about how Microsoft is ramping up to serve AI workloads later this month. Microsoft CTO Kevin Scott will be speaking at TechCrunch Disrupt, which will be held October 27 to October 29 in San Francisco.