2 Sources
2 Sources
[1]
Exclusive: Google deepens Thinking Machines Lab ties with new multi-billion-dollar deal | TechCrunch
Former OpenAI executive Mira Murati's startup, Thinking Machines Lab, has signed a new multi-billion-dollar agreement to expand its use of Google Cloud's AI infrastructure, including systems powered by Nvidia's latest GPUs, TechCrunch has exclusively learned. The deal is valued in the single-digit billions, according to a source familiar with the matter, and includes access to Google's latest AI systems built atop Nvidia's new GB300 chips, alongside infrastructure services to support model training and deployment. Google has been actively striking a number of cloud deals with AI developers as it aims to wrap together its cloud offerings with other services like storage, a Kubernetes engine, and Spanner, its database product. Earlier this month, Anthropic signed an agreement with Google and Broadcom for multiple gigawatts of tensor processing unit (TPUs) capacity (these are Google's custom-designed AI chips for machine learning workloads). But the competition is fierce. Just this week, Anthropic also signed a new agreement with Amazon to secure up to 5 gigawatts of capacity for training and deploying Claude. Earlier this year, Thinking Machines partnered with Nvidia in a deal that included an investment from the chipmaker. But this is the first time the lab has struck a deal with a cloud services provider. The deal is not exclusive, so Thinking Machines may use multiple cloud providers over time, but it's still a sign that Google is looking to lock in fast-growing frontier labs early. Murati left her job as OpenAI's chief technologist and founded Thinking Machines in February 2025. The company, which soon afterwards raised a $2 billion seed round at a $12 billion valuation, has remained highly secretive, but launched its first product in October. Dubbed Tinker, it's a tool that automates the creation of custom frontier AI models. Wednesday's deal provided some insight into what Thinking Machines is developing. In a press release, Google noted that it can support the startup's reinforcement learning workloads, which Tinker's architecture relies on. Reinforcement learning is a training approach that has underpinned recent breakthroughs at labs, including DeepMind and OpenAI, and the scale of the Google Cloud deal reflects how computationally expensive that work can get. Thinking Machines is among the first Google Cloud customers to access its GB300-powered systems, which offer a 2X improvement in training and serving speed compared to prior-generation GPUs, per Google. "Google Cloud got us running at record speed with the reliability we demand," Myle Ott, a founding researcher at Thinking Machines, said in a statement.
[2]
Google Cloud inks AI infrastructure deal with Thinking Machines - SiliconANGLE
Google Cloud inks AI infrastructure deal with Thinking Machines Thinking Machines Lab Inc. will expand its use of Google LLC's cloud platform as part of a partnership the companies announced this morning. TechCrunch cited a source as saying that the deal is valued in the "single-digit billions". Thinking Machines is an artificial intelligence startup led by Chief Executive Mira Murati, the former Chief Technology Officer of OpenAI Group PBC. It offers a cloud service called Tinker that enables developers to create custom versions of open-source large language models. Tinker performs customization by attaching add-ons to LLMs' core code. Thinking Machines will move some of its workloads to Google Cloud's A4X Max instances, which are specifically optimized for AI models. Each virtual machine provides access to 4 of Nvidia Corp.'s Blackwell Ultra graphics processing units. They're supported by 2 central processing units that feature 72 cores apiece. The chips run in liquid-cooled GB300 NVL72 appliances. The systems, which are made by Nvidia, come with 37 terabytes of memory and 130 terabits per second of internal bandwidth for moving data between processors. According to Google Cloud, Thinking Machines is one of the first customers to use its NVL72 infrastructure. The chips in A4X Max instances use a technology called RoCE to exchange data. It skips several of the steps usually involved in processing packets, which boosts throughput. The data travels over a network that features a so-called rail-aligned topology. Packets often have to travel through multiple network devices before reaching their destination. Each such device is known as a hop. A rail-aligned topology creates dedicated network links, or rails, between GPUs that minimize the number of hops data must go through, which boosts workload performance. Google uses Nvidia's ConnectX network interface cards, or NICs, to coordinate GPU traffic. A NIC is a chip that acts as an interface between a server and the data center network to which it's attached. The ConnectX chips run alongside Google's internally-developed Titanium NICs. A4X Max instances use the latter modules to process traffic between GPUs and external systems such as other Google Cloud services. Thinking Machines is already using several of those services to run its workloads. The company keeps information in Google Cloud Storage, the Spanner relational database and a custom cache. It relies on a fourth Google Cloud service called Cluster Director to automatically fix certain technical issues.
Share
Share
Copy Link
Mira Murati's Thinking Machines Lab has secured a multi-billion-dollar agreement with Google Cloud to expand its AI infrastructure access, including systems powered by Nvidia's latest GB300 chips. The deal marks the first cloud services partnership for the AI startup, which raised $2 billion at a $12 billion valuation earlier this year and launched its custom AI model tool, Tinker, in October.
Thinking Machines Lab, the AI startup founded by former OpenAI executive Mira Murati, has signed a new AI infrastructure deal with Google Cloud valued in the single-digit billions, according to a source familiar with the matter
1
. This multi-billion-dollar deal provides the company with access to Google's latest AI systems built on Nvidia's GB300 chips, alongside infrastructure services to support AI model training and deployment.
Source: TechCrunch
The agreement represents the first time Thinking Machines Lab has struck a deal with a cloud services provider, though the partnership is not exclusive, allowing the company to potentially work with multiple cloud providers over time
1
. Murati left her position as OpenAI's chief technologist and founded Thinking Machines in February 2025. The company raised a $2 billion seed round at a $12 billion valuation and launched its first product, Tinker, in October—a tool that automates the creation of custom frontier AI models1
.Thinking Machines will leverage Google Cloud's A4X Max instances, which are specifically optimized for AI workloads
2
. Each virtual machine provides access to 4 of Nvidia's Blackwell Ultra graphics processing units, supported by 2 central processing units featuring 72 cores apiece. The Nvidia GPUs run in liquid-cooled GB300 NVL72 appliances, which come with 37 terabytes of memory and 130 terabits per second of internal bandwidth2
. Thinking Machines is among the first Google Cloud customers to access its GB300-powered systems, which offer a 2X improvement in training and serving speed compared to prior-generation GPUs1
.
Source: SiliconANGLE
The cloud computing partnership provides critical insight into what Thinking Machines is developing. Google noted in its press release that it can support the startup's reinforcement learning workloads, which Tinker's architecture relies on
1
. Reinforcement learning is a training approach that has underpinned recent breakthroughs at labs including DeepMind and OpenAI, and the scale of this deal reflects how computationally expensive that work can get. Tinker enables developers to create custom versions of open-source large language models by attaching add-ons to LLMs' core code2
.Google has been actively striking cloud deals with AI developers, aiming to bundle its cloud offerings with services like storage, Kubernetes engine, and Spanner, its database product
1
. Earlier this month, Anthropic signed an agreement with Google and Broadcom for multiple gigawatts of TPUs capacity. However, competition among cloud providers remains fierce—Anthropic also signed a new agreement with Amazon to secure up to 5 gigawatts of capacity for training and deploying Claude1
. The deal signals Google's intent to lock in fast-growing frontier labs early, positioning itself as a preferred infrastructure partner.Related Stories
The A4X Max instances use RoCE technology to exchange data between chips, skipping several steps usually involved in processing packets to boost throughput
2
. Data travels over a network featuring a rail-aligned topology that creates dedicated network links between GPUs, minimizing the number of hops data must travel through and improving workload performance. Google uses Nvidia's ConnectX network interface cards alongside its internally-developed Titanium NICs to coordinate GPU traffic and process traffic between GPUs and external systems2
.Thinking Machines is already using several Google Cloud services to run its workloads, keeping information in Google Cloud Storage, the Spanner relational database, and a custom cache
2
. The company relies on Cluster Director, a Google Cloud service that automatically fixes certain technical issues. "Google Cloud got us running at record speed with the reliability we demand," said Myle Ott, a founding researcher at Thinking Machines1
. Earlier this year, Thinking Machines partnered with Nvidia in a deal that included an investment from the chipmaker, demonstrating the startup's strategy of building relationships across the AI infrastructure ecosystem.Summarized by
Navi
[1]
10 Mar 2026•Technology

Yesterday•Technology

22 Oct 2025•Technology

1
Policy and Regulation

2
Technology

3
Business and Economy
