OpenAI strikes $10 billion AI chip deal with Cerebras to power faster ChatGPT responses

Reviewed byNidhi Govil

10 Sources

Share

OpenAI has signed a multi-year agreement with AI chipmaker Cerebras worth over $10 billion to deliver 750 megawatts of computing power through 2028. The deal aims to accelerate ChatGPT's response times and support real-time AI applications, while reducing OpenAI's reliance on Nvidia's GPU-based systems. Cerebras will build and lease dedicated datacenters featuring its wafer-scale accelerators.

OpenAI and Cerebras Forge Massive Computing Partnership

OpenAI announced Wednesday that it reached a multi-year agreement with AI chipmaker Cerebras in a $10 billion deal that will deliver 750 megawatts of computing power starting this year and continuing through 2028

1

. The partnership marks a strategic move by OpenAI to diversify its supply chain beyond traditional GPU providers and accelerate inference capabilities for ChatGPT users.

Source: SiliconANGLE

Source: SiliconANGLE

According to sources familiar with the matter, the AI chip deal is valued at more than $10 billion, making it one of the largest compute agreements in the AI industry

3

. Cerebras will take on the risk of building and leasing datacenters equipped with its wafer-scale AI accelerators to serve OpenAI's inference workloads

2

. The AI infrastructure will be deployed in multiple stages, with Cerebras hosting the systems dedicated to delivering low-latency responses.

Why Speed Matters for Real-Time AI Applications

Both companies emphasized that the partnership centers on delivering faster outputs for OpenAI's customers. "Integrating Cerebras into our mix of compute solutions is all about making our AI respond much faster," OpenAI explained in a blog post

2

. When users ask complex questions, generate code, create images, or run AI agents, there's a loop happening behind the scenes where the model processes requests and sends responses back.

Andrew Feldman, co-founder and CEO of the semiconductor startup, drew a powerful analogy: "Just as broadband transformed the internet, real-time inference will transform AI"

1

. He emphasized that AI's inference stage—the process of getting AI models to respond to queries—is crucial to advancement, and that's where Cerebras' products excel. Recent work showed Cerebras running OpenAI's GPT-OSS-120B model 15 times faster than conventional hardware

3

.

Sachin Katti of OpenAI stated that "Cerebras adds a dedicated low-latency inference solution to our platform. That means faster responses, more natural interactions, and a stronger foundation to scale real-time AI to many more people"

1

. In the age of reasoning models and AI agents, faster inference means models can "think" for longer without compromising on interactivity

2

.

Technical Advantages of Cerebras' Wafer-Scale Architecture

By integrating Cerebras' wafer-scale compute architecture into its inference pipeline, OpenAI can leverage the chip's massive SRAM capacity to accelerate processing power. Each WSE-3 accelerator measures 46,225 mm² and is equipped with 44 GB of SRAM

2

. Compared to the HBM found on modern GPUs, SRAM operates several orders of magnitude faster. While a single Nvidia Rubin GPU can deliver around 22 TB/s of memory bandwidth, Cerebras' chips achieve nearly 1,000x that at 21 petabytes per second

2

.

Source: The Register

Source: The Register

Running models like OpenAI's gpt-oss 120B, Cerebras' chips can reportedly achieve single-user performance of 3,098 tokens per second compared to 885 tok/s for competitor Together AI, which uses Nvidia GPUs

2

. This performance advantage positions Cerebras as a compelling alternative in the market dominance currently held by Nvidia.

OpenAI's Strategy to Diversify Beyond Nvidia

The partnership represents OpenAI's continued effort to diversify supply chain dependencies as it pursues an aggressive expansion plan. "OpenAI's compute strategy is to build a resilient portfolio that matches the right systems to the right workloads," said Katti

1

. While Nvidia CEO Jensen Huang boasted in November that "everything that OpenAI does runs on Nvidia today," competition is clearly emerging

4

.

Last year, OpenAI committed more than $1.4 trillion to infrastructure deals with companies including Nvidia, Advanced Micro Devices, and Broadcom

4

. In September, Nvidia announced it would invest as much as $100 billion in OpenAI to build AI infrastructure with at least 10 gigawatts of power capacity. In October, AMD said it would deploy 6 gigawatts worth of graphics processing units over multiple years for OpenAI

3

. OpenAI is also developing its own chip with Broadcom

3

.

Implications for Cerebras' IPO and Market Position

For Cerebras, this high-profile win moves it closer to tapping into the tens of billions of dollars being poured into new AI infrastructure. The company has been around for over a decade, but its star has risen significantly since the launch of ChatGPT in 2022 and the AI boom that followed

1

. Feldman noted that "this transaction launches us into the big league and launches high-speed inference into the mainstream"

3

.

Source: ET

Source: ET

Cerebras filed for an IPO in September 2024, revealing that revenue in the second quarter approached $70 million, up from about $6 million in the second quarter of 2023

5

. However, the company withdrew the paperwork in October after announcing a $1.1 billion funding round that valued it at $8.1 billion. The company is now in talks to raise another billion dollars at a $22 billion valuation

1

. The OpenAI deal will help diversify Cerebras away from G42, which accounted for 87% of revenue in the first half of 2024

5

.

It's worth noting that Sam Altman is already an investor in Cerebras, and OpenAI once considered acquiring the company

1

. The relationship between the two companies dates back to 2017, when they first began exploring collaboration

3

. Greg Brockman, OpenAI co-founder and president, stated that "this partnership will make ChatGPT not just the most capable but also the fastest AI platform in the world," helping unlock "the next generation of use cases and onboard the next billion users to AI"

3

.

Today's Top Stories

TheOutpost.ai

Your Daily Dose of Curated AI News

Don’t drown in AI news. We cut through the noise - filtering, ranking and summarizing the most important AI news, breakthroughs and research daily. Spend less time searching for the latest in AI and get straight to action.

© 2026 Triveous Technologies Private Limited
Instagram logo
LinkedIn logo