Google Cloud signs multi-billion-dollar AI infrastructure deal with Thinking Machines Lab

2 Sources

Share

Mira Murati's Thinking Machines Lab has secured a multi-billion-dollar agreement with Google Cloud to expand its AI infrastructure access, including systems powered by Nvidia's latest GB300 chips. The deal marks the first cloud services partnership for the AI startup, which raised $2 billion at a $12 billion valuation earlier this year and launched its custom AI model tool, Tinker, in October.

Thinking Machines Lab Secures Major Google Cloud Partnership

Thinking Machines Lab, the AI startup founded by former OpenAI executive Mira Murati, has signed a new AI infrastructure deal with Google Cloud valued in the single-digit billions, according to a source familiar with the matter

1

. This multi-billion-dollar deal provides the company with access to Google's latest AI systems built on Nvidia's GB300 chips, alongside infrastructure services to support AI model training and deployment.

Source: TechCrunch

Source: TechCrunch

First Cloud Partnership for Murati's AI Startup

The agreement represents the first time Thinking Machines Lab has struck a deal with a cloud services provider, though the partnership is not exclusive, allowing the company to potentially work with multiple cloud providers over time

1

. Murati left her position as OpenAI's chief technologist and founded Thinking Machines in February 2025. The company raised a $2 billion seed round at a $12 billion valuation and launched its first product, Tinker, in October—a tool that automates the creation of custom frontier AI models

1

.

Access to Cutting-Edge AI-Optimized Infrastructure

Thinking Machines will leverage Google Cloud's A4X Max instances, which are specifically optimized for AI workloads

2

. Each virtual machine provides access to 4 of Nvidia's Blackwell Ultra graphics processing units, supported by 2 central processing units featuring 72 cores apiece. The Nvidia GPUs run in liquid-cooled GB300 NVL72 appliances, which come with 37 terabytes of memory and 130 terabits per second of internal bandwidth

2

. Thinking Machines is among the first Google Cloud customers to access its GB300-powered systems, which offer a 2X improvement in training and serving speed compared to prior-generation GPUs

1

.

Source: SiliconANGLE

Source: SiliconANGLE

Supporting Reinforcement Learning Workloads

The cloud computing partnership provides critical insight into what Thinking Machines is developing. Google noted in its press release that it can support the startup's reinforcement learning workloads, which Tinker's architecture relies on

1

. Reinforcement learning is a training approach that has underpinned recent breakthroughs at labs including DeepMind and OpenAI, and the scale of this deal reflects how computationally expensive that work can get. Tinker enables developers to create custom versions of open-source large language models by attaching add-ons to LLMs' core code

2

.

Google's Strategy to Lock in Frontier AI Labs

Google has been actively striking cloud deals with AI developers, aiming to bundle its cloud offerings with services like storage, Kubernetes engine, and Spanner, its database product

1

. Earlier this month, Anthropic signed an agreement with Google and Broadcom for multiple gigawatts of TPUs capacity. However, competition among cloud providers remains fierce—Anthropic also signed a new agreement with Amazon to secure up to 5 gigawatts of capacity for training and deploying Claude

1

. The deal signals Google's intent to lock in fast-growing frontier labs early, positioning itself as a preferred infrastructure partner.

Advanced Network Architecture Boosts Performance

The A4X Max instances use RoCE technology to exchange data between chips, skipping several steps usually involved in processing packets to boost throughput

2

. Data travels over a network featuring a rail-aligned topology that creates dedicated network links between GPUs, minimizing the number of hops data must travel through and improving workload performance. Google uses Nvidia's ConnectX network interface cards alongside its internally-developed Titanium NICs to coordinate GPU traffic and process traffic between GPUs and external systems

2

.

Thinking Machines Already Leveraging Google Cloud Services

Thinking Machines is already using several Google Cloud services to run its workloads, keeping information in Google Cloud Storage, the Spanner relational database, and a custom cache

2

. The company relies on Cluster Director, a Google Cloud service that automatically fixes certain technical issues. "Google Cloud got us running at record speed with the reliability we demand," said Myle Ott, a founding researcher at Thinking Machines

1

. Earlier this year, Thinking Machines partnered with Nvidia in a deal that included an investment from the chipmaker, demonstrating the startup's strategy of building relationships across the AI infrastructure ecosystem.

Today's Top Stories

TheOutpost.ai

Your Daily Dose of Curated AI News

Don’t drown in AI news. We cut through the noise - filtering, ranking and summarizing the most important AI news, breakthroughs and research daily. Spend less time searching for the latest in AI and get straight to action.

© 2026 Triveous Technologies Private Limited
Instagram logo
LinkedIn logo