NVIDIA declares AI inference the next boom as Jensen Huang unveils ambitious agent and robot strategy

Reviewed byNidhi Govil

4 Sources

Share

At GTC 2026, NVIDIA CEO Jensen Huang announced a strategic shift from training to AI inference, calling it the new inflection point. The company unveiled Vera Rubin, a platform promising 10x higher inference throughput, and revealed a $20 billion Groq licensing deal for low-latency inference technology. Huang positioned NVIDIA as builder of the entire AI infrastructure stack, from data centers to orbital facilities.

NVIDIA Shifts Focus to AI Inference as the Next Frontier

At the NVIDIA GTC keynote in San Jose, CEO Jensen Huang made a clear declaration: the center of gravity in artificial intelligence has shifted from training models to deploying them. "The inflection point for inference" has arrived, Huang told attendees during the two-hour presentation, marking what he believes will define the next phase of AI economics

1

. The message wasn't subtle. AI inference—the process of running trained models to generate outputs—is now where NVIDIA sees the biggest opportunity for growth and monetization

2

.

Source: Digit

Source: Digit

Huang's argument rested on a fundamental shift in how AI systems operate. "Computing used to be retrieval-based, now it's generative," he explained, positioning inference as the production layer where AI agents actually perform work rather than simply answer questions

4

. For NVIDIA, a company with a $4.4 trillion market cap—double what it was two years ago—this pivot addresses critical questions about whether the AI infrastructure boom can sustain itself beyond the training phase

1

.

Vera Rubin Platform Promises Massive Inference Performance Gains

The centerpiece of NVIDIA's AI strategy came in the form of Vera Rubin, a new platform that Huang described as "a generational leap" built around seven chips and five rack-scale systems

2

. NVIDIA claims the platform can train large mixture-of-experts models with one-fourth the number of GPUs versus Blackwell and deliver up to 10 times higher inference throughput per watt at one-tenth the cost per token. The company also previewed Feynman, the generation beyond Rubin, signaling that NVIDIA's roadmap extends well into the future

2

.

Source: Quartz

Source: Quartz

But Huang wasn't just selling faster chips. He unveiled a Vera Rubin DSX AI factory reference design, DSX simulation tools for planning AI factories before construction, and a broader menu of storage, networking, and system components designed to operate as vertically integrated AI infrastructure

2

. The message was unmistakable: NVIDIA wants customers to think about campuses, not servers, and to treat AI infrastructure like a utility

2

.

Huang even took the vision orbital, announcing plans for Vera Rubin-based systems aimed at data centers in space for autonomous space operations, though no timeline was provided

1

2

.

The $20 Billion Groq Deal and the Inference Land Grab

Perhaps the most strategically significant announcement came in the form of NVIDIA's non-exclusive $20 billion licensing agreement with Groq for low-latency inference technology

3

. On NVIDIA's last earnings call, Jensen Huang explicitly compared the Groq deal to the company's acquisition of Mellanox, suggesting it will "extend NVIDIA's architecture with Groq's innovations" the same way Mellanox extended networking capabilities

3

.

The Groq technology addresses a glaring gap in NVIDIA's portfolio: low-latency inference at edge computing environments, where agentic AI systems need to deliver value with minimal delay . By integrating Groq's decoder technology into the CUDA ecosystem, NVIDIA aims to collapse one of the highest-value inference opportunities back into its control plane

3

. Industry observers see this as NVIDIA's attempt to dominate the inference land grab before competitors can establish footholds in the fragmented market

3

.

AI Agents Take Center Stage with New Toolkit and Models

"Claude Code and OpenClaw have sparked the agent inflection point," Huang declared, positioning AI agents as systems that execute tasks rather than simply respond to queries

1

. NVIDIA introduced Nemo Claw, a protective layer of security and stability for OpenClaw (formerly Clawdbot), now available in preview

1

. The company also launched an expansive NVIDIA AI Agent Toolkit designed to help enterprises build their own models, along with the AI-Q blueprint that claims to cut query costs by more than 50% through a hybrid mix of frontier and NVIDIA's own open models

2

.

Huang described agentic AI as a fundamental shift: "For the first time, you don't ask AI what, where, when, how. You ask it to create, do, build. It's able to solve problems and actually perform tasks"

4

. NVIDIA also unveiled the Nemotron 3 omni-understanding model for complex reasoning, and announced the Nemotron Coalition with partners including Black Forest Labs, Cursor, LangChain, Mistral, and Perplexity

1

2

.

Physical AI and Robots Signal Long-Term Infrastructure Bet

Huang previewed GR00T N2, a next-generation robot foundation model based on DreamZero research that NVIDIA claims more than doubles success versus leading VLA models on new tasks in new environments

2

. The physical AI segment featured 110 AI-powered robots from NVIDIA-partnered companies, including a somewhat awkward interaction with an Olaf robot from Disney's Frozen

1

.

Source: Mashable

Source: Mashable

Industry analysts suggest this physical AI push may be the most strategically important announcement from the NVIDIA GTC keynote. While chatbots sparked initial investor excitement, robots, industrial systems, and autonomous machines require endless training data, simulation, networking, sensors, and edge compute—potentially sustaining the infrastructure spending boom for years

2

.

Enterprise AI Built on Structured Data and Token Factories

Huang made a compelling case for why enterprise AI requires more than generative capabilities. "Structured data is the foundation of trustworthy AI," he stated, emphasizing that businesses need AI systems that plug into existing data, records, processes, and governance rules without hallucinating

4

. He noted that approximately 90% of data generated annually is unstructured and largely unused, representing a massive opportunity for AI systems that can extract meaning and embed it into searchable structures

4

.

Huang introduced a new framing for data centers: "Your data centre, it used to be a data centre for files. It's now a factory to generate tokens"

4

. This concept of token factories positions NVIDIA's AI strategy around measuring infrastructure by cost per token and usefulness per watt, treating AI deployment as an industrial production system

4

.

Mixed Reception for Consumer-Facing Announcements

Not all announcements landed well. NVIDIA revealed DLSS 5, the next iteration of its AI upscaling software coming this fall, which the company describes as a "breakthrough in visual fidelity" that "infuses pixels with photorealistic lighting and materials"

1

. Huang showcased comparisons using Resident Evil: Requiem, Hogwarts Legacy, and Starfield. However, gamers took to social media frustrated that NVIDIA was attempting to "fix" games like Resident Evil: Requiem, which had already won praise for its graphics

1

.

The keynote concluded with AI-generated country music featuring Huang's avatar and robots around a campfire, leaving the audience with two possible futures: one where AI agents help launch NVIDIA into the stratosphere, and another where agents deliver little ROI and the company's market position weakens

1

. For now, NVIDIA's bet on the next AI boom belonging to inference represents a calculated attempt to justify its massive valuation and define the infrastructure layer for the next decade of AI development.

Today's Top Stories

TheOutpost.ai

Your Daily Dose of Curated AI News

Don’t drown in AI news. We cut through the noise - filtering, ranking and summarizing the most important AI news, breakthroughs and research daily. Spend less time searching for the latest in AI and get straight to action.

© 2026 Triveous Technologies Private Limited
Instagram logo
LinkedIn logo