Google builds specialized inference AI chips with four partners to challenge Nvidia dominance

Reviewed byNidhi Govil

5 Sources

Share

Google is assembling the AI industry's most diversified custom chip supply chain, partnering with Broadcom, MediaTek, Marvell, and Intel to develop specialized inference chips. The move positions Google's Tensor Processing Units as a direct challenge to Nvidia's dominance as the battleground shifts from training to inference workloads, where cost per query determines AI business economics.

Google Shifts Focus to Specialized Inference AI Chips

Google is developing new AI chips dedicated to inference workloads, marking a strategic shift that positions the company to directly challenge Nvidia in the fastest-growing segment of the AI semiconductor market. After months of surging demand for its Tensor Processing Units (TPUs), Google plans to announce its next-generation custom-designed chips at the Google Cloud Next conference in Las Vegas this week

1

. The company's Chief Scientist Jeff Dean explained that as demand grows for quickly processing AI queries, "it now becomes sensible to specialize chips more for training or more for inference workloads"

3

.

Source: Market Screener

Source: Market Screener

Inference represents the stage where AI models actually perform their jobs—fielding queries and producing outputs after training is complete. While Google has previously touted inference capabilities for its chips, the company initially resisted releasing separate chips for training and inference

1

. That approach is changing as the AI spending boom moves from training to inference, with Gartner analyst Chirag Dekate noting that "the battleground is shifting towards inference"

2

.

Building the Most Diversified Custom Chip Supply Chain

Google is assembling the AI industry's most diversified custom chip supply chain, involving four distinct design partners: Broadcom, MediaTek, Marvell, and Intel

2

. The TPUv8 strategy splits the next generation explicitly, with Broadcom's "Sunfish" chip handling training workloads and MediaTek's "Zebrafish" chip targeting cost-efficient inference at 20-30% lower cost

4

. Both chips target TSMC's 2-nanometre process node for deployment in late 2027

2

.

Source: Wccftech

Source: Wccftech

Broadcom, which signed a long-term agreement on April 6 to supply TPUs and networking components through 2031, commands more than 70% of the custom AI accelerator market and projects $100 billion in AI chip revenue by 2027

2

. MediaTek's involvement began with I/O modules and peripheral components on Ironwood, Google's seventh-generation TPU, where its designs run 20 to 30% cheaper than alternatives

2

. Marvell Technology is in talks with Google to develop a memory processing unit and a new inference-focused TPU, with plans to produce nearly two million of the memory processing units

2

.

Google's Infrastructure Advantage in AI Inference Workloads

Google brings unique strengths to the competitive landscape challenging Nvidia, including a decade of experience designing chips, vast resources from its online search profits, and firsthand insights on AI models

1

. Among top AI developers, only Google makes its own chips at significant scale, allowing it to share vital feedback between teams to better customize hardware. Demis Hassabis, CEO of Google DeepMind, told Bloomberg that interest in TPUs is particularly high from leading AI labs

1

.

Source: Bloomberg

Source: Bloomberg

The current Ironwood TPU delivers ten times the peak performance of the TPU v5p, offers 192 gigabytes of HBM3E memory per chip with 7.2 terabytes per second of bandwidth, and scales to 9,216 liquid-cooled chips in a single superpod producing 42.5 FP8 exaflops

2

. Google plans to produce millions of units this year, with Anthropic committing to up to 1 million TPUs under an expanded agreement unveiled in October

2

. Meta Platforms also signed a multibillion-dollar deal to use TPUs through Google Cloud over several years, with Santosh Janardhan, Meta's head of infrastructure, noting that "it does look like there might be inference advantages"

1

.

Why Inference Economics Matter for Google's AI Infrastructure

The shift from training to inference as the dominant AI compute cost is the strategic premise behind Google's entire chip programme. Training a frontier model is a singular, intensive event, while inference is continuous and scales with every user, every query, and every product that incorporates AI

2

. Google serves billions of AI-augmented search queries, Gemini conversations, and Cloud AI API calls daily. At that scale, the cost per inference determines the economics of the entire AI business.

Natalie Serrino, co-founder at Gimlet Labs, a startup that makes software for routing AI tasks to the best chip for each job, said today's TPUs are a strong choice for processing results for the emerging crop of AI agents that field more complex work on a user's behalf: "They are very good tools for the workload that is exploding"

1

. Nvidia's GPUs remain dominant for training workloads, where programmability and the CUDA software ecosystem create switching costs that custom chips cannot easily replicate

2

. However, inference workloads are more predictable, more repetitive, and more amenable to the kind of fixed-function optimization that custom silicon excels at.

While Nvidia CEO Jensen Huang stressed at the company's GTC conference that its chips can handle applications "you can't do with TPUs," Google uses both TPUs and GPUs for its own AI projects

3

. Supply constraints may complicate Google's ambitions, with an unnamed startup executive describing chip scarcity as a real obstacle

3

. However, available supply is being steered toward leading AI organizations, according to Hassabis, who described them as "the more elite teams"

3

. The upcoming TPU family is expected to boost semiconductor and assembly supply chain markets while creating upgrade opportunities for peripheral components, including all-optical switches, liquid cooling, power supplies, and optical communications companies

4

.

Today's Top Stories

TheOutpost.ai

Your Daily Dose of Curated AI News

Don’t drown in AI news. We cut through the noise - filtering, ranking and summarizing the most important AI news, breakthroughs and research daily. Spend less time searching for the latest in AI and get straight to action.

© 2026 Triveous Technologies Private Limited
Instagram logo
LinkedIn logo