Nvidia launches Vera Rubin platform at CES 2026, promising 10x cost reduction for AI computing

Reviewed byNidhi Govil

26 Sources

Share

Nvidia CEO Jensen Huang announced the Vera Rubin AI computing platform is in full production at CES 2026. The next-generation AI platform delivers five times faster AI inference than Blackwell architecture while cutting costs by up to 10x. Major cloud providers including Microsoft, AWS, and Google Cloud will deploy Rubin systems starting in the second half of 2026.

Nvidia Unveils Vera Rubin Platform in Full Production

Nvidia CEO Jensen Huang officially launched the company's Vera Rubin chip architecture at CES 2026, declaring the next-generation AI platform is already in full production

1

. The announcement marks a shift in how Nvidia positions its products, moving away from individual GPU sales toward complete rack-scale AI systems designed to address the skyrocketing computational demands of modern AI models

5

.

Source: Tom's Hardware

Source: Tom's Hardware

"Vera Rubin is designed to address this fundamental challenge that we have: The amount of computation necessary for AI is skyrocketing," Huang told the audience at the Consumer Electronics show

1

. The Rubin platform will replace the Blackwell architecture as Nvidia's flagship AI computing solution, with production expected to ramp up further in the second half of 2026.

Dramatic Performance Gains Over Blackwell Architecture

The Rubin platform delivers substantial improvements in both speed and cost efficiency compared to its predecessor. According to Nvidia's tests, the Rubin GPU operates three and a half times faster than Blackwell on model training tasks and five times faster on AI inference tasks, reaching as high as 50 petaflops of NVFP4 computational power

1

4

. The platform can train large mixture-of-experts AI models using roughly one-fourth as many chips as Blackwell requires while delivering up to a 10x reduction in inference token costs

2

4

.

Power efficiency represents another major advance, with the new platform supporting eight times more inference compute per watt

1

. These gains could make advanced AI systems significantly cheaper to operate and make it harder for Nvidia's customers to justify moving away from its hardware, analysts note

2

.

Six-Chip Architecture Addresses AI Bottlenecks

Named after astronomer Vera Florence Cooper Rubin, the architecture consists of six separate chips designed to work in concert

1

. At the center sits the Rubin GPU, but the system also includes a Vera CPU built with 88 custom Olympus cores designed specifically for agentic reasoning

4

. Both the Rubin GPU and Vera CPU are manufactured using Taiwan Semiconductor Manufacturing Company's 3 nanometer fabrication process with the most advanced bandwidth memory technology currently available

2

.

Source: Wccftech

Source: Wccftech

The architecture addresses growing bottlenecks in storage and interconnection through improvements in the BlueField DPU and NVLink systems. Dion Harris, Nvidia's senior director of AI infrastructure solutions, explained that new workflows like agentic AI place significant stress on KV cache memory systems. "We've introduced a new tier of storage that connects externally to the compute device, which allows you to scale your storage pool much more efficiently," Harris told reporters

1

. The BlueField-4 DPU introduces a shared memory tier for long-context inference, treating context as a first-class system resource rather than a per-GPU issue

5

.

Nvidia's sixth-generation NVLink interconnects and Spectrum-6 Ethernet switch provide the networking backbone, while ConnectX-9 SuperNIC handles high-speed networking

4

.

Major Cloud Providers Commit to Deployment

Rubin chips are already slated for use by nearly every major cloud provider. Microsoft and CoreWeave will be among the first companies to begin offering services powered by Rubin systems later this year

2

. Two major AI data centers that Microsoft is currently building in Georgia and Wisconsin will eventually include thousands of Rubin chips

2

. Other confirmed partners include Amazon Web Services, Google Cloud, Anthropic, and OpenAI

1

4

.

Rubin systems will also power HPE's Blue Lion supercomputer and the upcoming Doudna supercomputer at Lawrence Berkeley National Lab

1

. Nvidia is also working with Red Hat to offer more products that will run on the new system for banks, automakers, airlines, and government agencies

2

.

Strategic Shift Toward System-Level Solutions

The Rubin platform launch signals a fundamental shift in Nvidia's business strategy. For the first time in roughly five years, Nvidia stood on the CES stage without a new consumer GPU announcement

5

. The company is no longer content to sell accelerators one card at a time; it is selling entire AI systems instead, reflecting how hyperscalers and AI labs now deploy hardware in standardized blocks measured in racks or data halls

5

.

Source: Tom's Hardware

Source: Tom's Hardware

The flagship Nvidia Vera Rubin NVL72 configuration combines 36 Nvidia Vera CPUs, 72 Nvidia Rubin GPUs, NVLink 6 switches, multiple ConnectX-9 SuperNICs, and BlueField-4 DPUs into a single logical system

4

. This emphasis on pre-integrated systems shortens deployment timelines and reduces the tuning work customers must do themselves

5

.

The launch comes amid intense competition to build AI infrastructure. On an earnings call in October 2025, Huang estimated that between $3 trillion and $4 trillion will be spent on AI infrastructure over the next five years

1

. Nvidia recently reported record high data center revenue, up 66 percent over the prior year, driven by demand for Blackwell and Blackwell Ultra GPUs

3

. The goal with the Rubin platform is to accelerate mainstream adoption of advanced large language models, particularly in the consumer space, by sharply reducing the astronomical costs that have held back widespread AI deployment

4

.

Today's Top Stories

TheOutpost.ai

Your Daily Dose of Curated AI News

Don’t drown in AI news. We cut through the noise - filtering, ranking and summarizing the most important AI news, breakthroughs and research daily. Spend less time searching for the latest in AI and get straight to action.

© 2026 Triveous Technologies Private Limited
Instagram logo
LinkedIn logo