Google taps Marvell Technology for new AI inference chips as compute costs shift to serving models

Reviewed byNidhi Govil

3 Sources

Share

Google is in talks with Marvell Technology to develop two new AI chips aimed at running models more efficiently. The discussions center on a memory processing unit designed to work alongside existing TPUs and a new inference-optimized TPU. This move adds Marvell as a third design partner alongside Broadcom and MediaTek, reflecting Google's strategy to diversify its custom silicon supply chain as inference becomes the dominant AI compute expense.

Google Expands Custom Silicon Partners with Marvell Technology

Google is in discussions with Marvell Technology to develop two new AI chips designed to run AI models more efficiently, according to reports from The Information

1

. The talks center on a memory processing unit that would work alongside Google's existing Tensor Processing Unit (TPU) infrastructure and a new inference-optimized TPU built specifically for serving AI models to users

2

. While the discussions have not yet produced a signed contract, the companies aim to finalize the memory processing unit design as soon as next year before handing it off for test production

1

.

Source: Wccftech

Source: Wccftech

Marvell would act in a design-services role, similar to MediaTek's involvement on Google's latest Ironwood TPU

2

. The timing is notable, coming just days after Broadcom, Google's primary custom chip partner, announced a long-term agreement to design and supply TPUs and networking components through 2031. This suggests Google is not replacing Broadcom but adding a third design partner to diversify its custom silicon supply chain

2

.

AI Inference Drives Strategic Shift in Chip Development

The push toward specialized AI inference chips reflects a fundamental shift in where compute costs accumulate. Training a frontier model is a one-time event requiring enormous compute for weeks or months, but AI inference runs continuously, serving every query from every user

2

. As AI products reach hundreds of millions of users, inference becomes the dominant expense, and purpose-built inference silicon delivers competitive advantages that general-purpose GPUs cannot match on cost or efficiency.

Google's seventh-generation TPU, Ironwood, debuted this month as what the company calls "the first Google TPU for the age of inference." It delivers ten times the peak performance of the TPU v5p and scales to 9,216 liquid-cooled chips in a Superpod spanning roughly 10 megawatts, producing 42.5 FP8 exaflops

2

. Google plans to build millions of Ironwood units this year. The Marvell-designed chips would supplement rather than replace Ironwood, potentially targeting different workload profiles or cost points.

Memory Processing Unit Addresses Critical Bottleneck

The first chip under discussion is a memory processing unit designed to pair with TPUs, potentially using in-memory processing to offset some memory requirements from the chip or system

3

. This approach could accelerate the memory subsystem for faster AI model performance, especially in the inferencing segment where memory bandwidth often becomes a constraint

3

.

Multi-Supplier Strategy Reduces Supply Chain Risk

Google has been pushing to make its TPUs a viable alternative to compete with Nvidia's GPUs, with TPU sales becoming a key driver of growth in Google's cloud revenue as it aims to show investors that AI investments are generating returns

1

. The approach mirrors how automotive companies manage component suppliers: no single vendor gets enough leverage to dictate terms

2

.

The custom silicon market context makes this strategy critical. The custom ASIC market is projected to grow 45% in 2026 and reach $118 billion by 2033

2

. Marvell's data centre revenue reached a record $6.1 billion in its fiscal year ending February 2026, with total revenue of $8.2 billion, up 42% year over year

2

. The company runs a custom silicon business with a $1.5 billion annual run rate across 18 cloud-provider design wins, building chips for Amazon, Microsoft, and Meta

2

.

Nvidia Investment Positions Marvell at Ecosystem Intersection

Nvidia invested $2 billion in Marvell at the end of March, partnering through NVLink Fusion to integrate Marvell's custom chips and networking with Nvidia's interconnect technology

2

. This positions Marvell at the intersection of both the GPU and ASIC ecosystems. In December 2025, Marvell acquired Celestial AI for up to $5.5 billion, gaining photonic interconnect technology that CEO Matt Murphy said would deliver "the industry's most complete connectivity platform for AI and cloud customers"

2

. Murphy is targeting 20% market share in the AI accelerator market and expects roughly 30% year-over-year revenue growth in fiscal 2027.

The Google-Marvell relationship has deeper roots than recent headlines suggest. The Information reported in 2023 that Google had been working since 2022 on a chip codenamed "Granite Redux" that would use Marvell instead of Broadcom, with Google expecting to save billions of dollars annually

2

. What changed between 2023 and now is that Google appears to have abandoned the idea of dropping Broadcom entirely, instead building a multi-supplier architecture in which Broadcom, MediaTek, and potentially Marvell each handle different parts of the TPU programme. As hyperscaler compute demands intensify and new AI chips enter production, watch for how Google balances performance requirements against supply chain risk while maintaining cost advantages over Nvidia's dominant position.

Today's Top Stories

TheOutpost.ai

Your Daily Dose of Curated AI News

Don’t drown in AI news. We cut through the noise - filtering, ranking and summarizing the most important AI news, breakthroughs and research daily. Spend less time searching for the latest in AI and get straight to action.

© 2026 Triveous Technologies Private Limited
Instagram logo
LinkedIn logo