Microsoft deploys Maia 200 AI chip but won't stop buying from Nvidia, Nadella confirms

Reviewed byNidhi Govil

27 Sources

Share

Microsoft has begun deploying its second-generation Maia 200 AI chip in Azure data centers, claiming 30% better performance per dollar and superior specs compared to Amazon and Google's custom silicon. Despite this milestone, CEO Satya Nadella confirmed the company will continue purchasing AI hardware from Nvidia and AMD, emphasizing the value of multi-vendor partnerships in the rapidly evolving AI infrastructure landscape.

Microsoft Deploys Maia 200 AI Chip in Azure Data Centers

Microsoft has started deploying its in-house AI chip, the Maia 200, in its Azure data centers this week, marking a significant step in the company's effort to reduce AI operational costs and compete with custom silicon from tech giants

1

. Built on TSMC's 3nm process, the chip packs more than 100 billion transistors and delivers over 10 petaflops of FP4 performance, positioning it as what Microsoft calls an "AI inference powerhouse"

2

. The Maia 200 represents the successor to Microsoft's first-generation Maia 100 released in 2023, with substantial improvements designed specifically for AI inference workloads rather than training

3

.

Source: Interesting Engineering

Source: Interesting Engineering

The chip's technical specifications reveal a focus on efficiency and scale. With 216GB of HBM3e memory providing 7TB/s of bandwidth and 272MB of on-die SRAM, the Maia 200 can handle large AI models with room for even bigger models in the future

4

. Microsoft claims the chip delivers 30% better performance per dollar than the latest generation AI hardware in its fleet, operating at 750 watts—nearly half the power consumption of Nvidia's Blackwell B300 Ultra

5

.

Outperforming Amazon and Google's Custom Silicon

Microsoft has positioned the Maia 200 to directly compete with Amazon's Trainium and Google's TPU offerings. The company claims its chip delivers three times the FP4 performance of third-generation Amazon Trainium chips and FP8 performance above Google's seventh-generation TPU

2

. This performance advantage matters as cloud providers race to reduce reliance on Nvidia and offer more cost-effective AI infrastructure to customers. The chip's design prioritizes inference—the compute-intensive work of running AI models in production—which has become an increasingly important cost factor as AI companies mature

2

.

The Maia 200's architecture includes specialized tile tensor units supporting FP8, FP6, and FP4 datatypes in hardware, though this comes with trade-offs for workloads requiring higher precision

5

. Microsoft has equipped each chip with 2.8TB/s of bidirectional bandwidth, enabling clusters of up to 6,144 chips that can pool compute and memory resources—totaling 61 exaflops of AI compute and 1.3 petabytes of HBM3e memory

5

.

Source: Motley Fool

Source: Motley Fool

Nadella Confirms Continued Nvidia and AMD Partnerships

Despite launching its own state-of-the-art AI chip, Microsoft CEO Satya Nadella made clear the company won't abandon its hardware partnerships. "We have a great partnership with Nvidia, with AMD. They are innovating. We are innovating," Nadella explained, adding that "you have to be ahead for all time to come"

1

. He emphasized that vertical integration doesn't mean building systems exclusively from in-house components, signaling Microsoft's strategy to maintain a heterogeneous AI hardware approach across its Azure data centers

4

.

Source: Observer

Source: Observer

This multi-vendor strategy reflects the ongoing supply constraints for cutting-edge AI hardware and the rapid pace of innovation in the sector. With Nvidia's Rubin chips promising a 5x uplift in inference performance compared to Blackwell or Maia 200 launching later this year, Microsoft's approach hedges against being locked into any single technology path

5

.

Superintelligence Team Gets First Access

Microsoft's Superintelligence team, led by former Google DeepMind co-founder Mustafa Suleyman, will be the first to use the Maia 200 as they develop frontier AI models

1

. This priority access signals Microsoft's ambition to build its own large AI models and potentially reduce dependence on OpenAI, Anthropic, and other model makers. The chip will also support OpenAI's models running on Azure, including hosting OpenAI's GPT-5.2 model and powering Microsoft 365 Copilot

3

.

Microsoft has invited developers, academics, frontier AI labs, and open-source contributors to an early preview of the Maia 200 software development kit

2

. The company is deploying these chips initially in its Azure US Central data center region, with additional rollouts planned for US West 3 in Phoenix, Arizona, and other locations as supply increases

4

. The Maia 200's development faced delays—originally codenamed Braga, it was meant for a 2025 release but arrived later than planned

4

.

Today's Top Stories

TheOutpost.ai

Your Daily Dose of Curated AI News

Don’t drown in AI news. We cut through the noise - filtering, ranking and summarizing the most important AI news, breakthroughs and research daily. Spend less time searching for the latest in AI and get straight to action.

© 2026 Triveous Technologies Private Limited
Instagram logo
LinkedIn logo