Nvidia launches Nemotron 3 open source AI models as Meta steps back from transparency

Reviewed byNidhi Govil

16 Sources

Share

Nvidia released its Nemotron 3 family of models in three sizes—Nano (30B), Super (100B), and Ultra (500B parameters)—with full training data transparency. The move positions the chip giant as a major force in open source AI just as Meta signals a retreat from openness, leaving Chinese models to dominate the space.

Nvidia Takes Leadership Position in Open Source AI

Nvidia has stepped decisively into the open source AI arena with the release of its Nemotron 3 family of models, positioning itself as a transparency leader at a moment when American tech giants are retreating from openness

1

. The chip giant, which has built its fortune supplying GPUs to AI companies, is now hedging against the possibility that firms like OpenAI, Google, and Anthropic might develop their own chips and move away from Nvidia's technology over time

1

.

The Nemotron 3 family of models arrives in three sizes: Nano with 30 billion parameters, Super with 100 billion, and Ultra with 500 billion

1

. The Nano model, available now on Hugging Face, increases throughput in tokens per second by four times and extends the context window to one million tokens—seven times larger than its predecessor

2

. Super is expected in January, while Ultra will arrive in March or April

2

.

Source: Digit

Source: Digit

Enterprise AI Adoption Drives Model Customization Focus

Nvidia is taking a more transparent approach than many US rivals by releasing the training data used to train Nemotron, a fact that should help engineers with model customization more easily

1

. The company is also releasing tools including a new hybrid latent mixture-of-experts model architecture, which Nvidia says excels at building AI agents that can take actions on computers or the web

1

.

Kari Briski, vice president of generative AI software for enterprise at Nvidia, explained that open AI models matter to builders for three key reasons: the growing need to customize models for particular tasks, the benefit of routing queries to different models, and the ability to extract more intelligent responses through simulated reasoning

1

. For enterprise AI adoption, exposing sensitive customer data or intellectual property to APIs for closed models like ChatGPT remains a non-starter

3

.

The company is releasing reinforcement learning libraries and training environments called NeMo Gym, enabling users to train AI agents through trial and error with rewards and punishments

3

. Jensen Huang stated that "open innovation is the foundation of AI progress," emphasizing how Nemotron transforms advanced AI into an open platform that gives developers the transparency and efficiency needed to build agentic systems at scale

1

.

Source: SiliconANGLE

Source: SiliconANGLE

Meta's Retreat Creates Opening in AI Development Ecosystem

The timing of Nvidia's push is strategic. Meta, which released the first advanced open models under the name Llama in February 2023, has signaled that future releases might not be open source

1

. Meta's rollout of Llama 4 in April 2025 received mediocre reviews, and Llama models no longer appear in the top 100 on LMSYS's LMArena Leaderboard, which is dominated by proprietary models from Google, xAI, Anthropic, and OpenAI, along with open source AI models from DeepSeek AI, Alibaba's Qwen, and Singapore-based Moonshot AI

2

.

A forthcoming Meta project code-named Avocado may launch as a closed model that Meta can sell access to, marking the biggest departure from the open source strategy Meta has promoted for years

2

. Meta's Chief AI Officer Alexandr Wang, installed this year, is an advocate of closed models

2

. When asked about Menlo Ventures' claim that open source is struggling, Briski replied: "I agree about the decline of Llama, but I don't agree with the decline of open source"

2

.

Chinese AI Models Dominate as American Firms Turn Proprietary

With American tech giants drawing back from open source, Chinese AI models have dominated the space with high-profile releases from DeepSeek and Alibaba

5

. Open models from Chinese companies are currently much more popular than those from US firms, according to data from Hugging Face

1

. Even OpenAI, which purports to be an "open" organization, unveiled only its first open source model in five years this past August with gpt-oss-120b and gpt-oss-20b

5

.

At Nvidia's GTC conference in Washington DC last October, Jensen Huang warned attendees of China's open source dominance, arguing that because China creates substantial open source software, Americans risk being "ill-prepared" if Chinese software "permeates the world"

5

. Briski emphasized that "this year alone, we had the most contributions and repositories on Hugging Face," signaling Nvidia's growing prominence

2

.

Technical Architecture Targets AI Infrastructure Needs

The Nemotron 3 models employ a novel hybrid latent mixture-of-experts architecture designed to minimize performance losses when processing long input sequences

3

. This combines Mamba-2 and Transformer architectures throughout the model's layers, with Mamba-2 providing efficiency for long sequences while transformers maintain precise reasoning and prevent context loss

3

. All models natively support a million token context window, equivalent to roughly 3,000 double-spaced pages

3

.

The mixture-of-experts architecture means only a fraction of total parameters activates for each token processed. Nemotron 3 Nano has 30 billion parameters but only 3 billion activate per token, putting less pressure on memory and delivering faster throughput than equivalent dense models

3

. The larger Super and Ultra models use Nvidia's NVFP4 data type and a latent MoE architecture where experts operate on shared latent representation, allowing models to call on 4x more experts at the same inference cost

3

.

Strategic Positioning for Hardware Dominance

Nvidia sees Nemotron as a tool to drive further reliance on its hardware empire. By distributing open models to developers worldwide, the company aims to ensure that new foundation models built will align best with its silicon rather than rapidly emerging Chinese alternatives

5

. "When we're the best development platform, people are going to choose us, choose our platform, choose our GPU, in order to be able to not just build today, but also build for tomorrow," Briski said

5

.

The 30-billion-parameter Nemotron 3 Nano runs efficiently on enterprise hardware like Nvidia's L40S or RTX Pro 6000 Server Edition, though 4-bit quantized versions should work on GPUs with as little as 24GB of video memory

3

. According to Artificial Analysis, the model delivers performance on par with gpt-oss-20B or Qwen3 models while offering greater flexibility for customization

3

. Nvidia is positioning these offerings as business-ready infrastructure for enterprises to build domain-specific AI agents without needing to create foundation models from scratch

4

.

Source: The Register

Source: The Register

Today's Top Stories

TheOutpost.ai

Your Daily Dose of Curated AI News

Don’t drown in AI news. We cut through the noise - filtering, ranking and summarizing the most important AI news, breakthroughs and research daily. Spend less time searching for the latest in AI and get straight to action.

© 2026 Triveous Technologies Private Limited
Instagram logo
LinkedIn logo