Amazon unveils AI Factories with Nvidia partnership and launches Trainium3 chip for on-premises AI

Reviewed byNidhi Govil

21 Sources

Share

AWS introduced AI Factories, a fully managed on-premises AI solution built with Nvidia, alongside its new Trainium3 chip that delivers four times more performance than its predecessor. The announcements signal Amazon's dual strategy to accelerate cloud computing capacity while addressing data sovereignty concerns for enterprises and governments requiring local AI infrastructure.

AWS AI Factories Bring On-Premises AI to Enterprise Datacenters

Amazon Web Services announced AWS AI Factories at its re:Invent 2025 conference, a fully managed solution that allows corporations and governments to run AI systems within their own data centers . The service addresses data sovereignty concerns by keeping sensitive information on-premises while AWS handles hardware installation and management. Customers provide the physical space and power, while AWS supplies the AI infrastructure, including compute, storage, and database services that operate like a private AWS Region

4

.

This collaboration with Nvidia enables customers to choose between Nvidia Blackwell GPUs or Amazon's new Trainium3 AI chip, combined with AWS networking, storage, and security tools . The service integrates with Amazon Bedrock for model selection and SageMaker for AI training and inference, providing a comprehensive platform for AI workloads without requiring customers to acquire or install hardware themselves

4

. AWS CEO Matt Garman revealed the concept originated from work with Saudi Arabia's Humain to build an "AI Zone" featuring up to 150,000 AI chips

4

.

Source: CRN

Source: CRN

Trainium3 AI Chip Delivers Major Performance Leap

AWS formally launched its Trainium3 UltraServer, powered by state-of-the-art 3-nanometer Trainium3 chip technology

2

. The third-generation system delivers four times more compute performance, memory bandwidth, and energy efficiency compared to previous generations

5

. Each UltraServer hosts 144 chips, and thousands can be linked together to provide up to 1 million Trainium3 chips—ten times the capacity of the previous generation

2

.

Source: Digit

Source: Digit

The Trainium3 accelerator features dual chiplets equipped with 144 GB of HBM3E memory and peak bandwidth of 4.9 TB/s

3

. With FP8 performance reaching 2,517 MXFP8 TFLOPS, the Trn3 UltraServer packs 0.36 ExaFLOPS of FP8 performance across its 144-chip configuration, matching Nvidia's NVL72 GB300 rack-scale AI systems

3

. Early customers including Anthropic, Japan's Karakuri, and Decart have already reduced AI training and inference costs by up to 50% using the new accelerators

2

5

.

Trainium4 Roadmap Signals Deeper Nvidia Integration

AWS teased its next-generation Trainium4 chip, already in development, which will support NVLink Fusion—Nvidia's high-speed chip interconnect technology

2

. This integration will allow Trainium4-powered systems to interoperate with Nvidia GPUs while leveraging Amazon's lower-cost server rack technology

2

. The move could make it easier to attract AI applications built with Nvidia's CUDA platform, which has become the de facto standard for AI workloads

2

.

Source: Market Screener

Source: Market Screener

Cloud Computing Capacity Race Intensifies

While AWS AI innovations capture attention, Wall Street analysts emphasize that cloud computing capacity expansion matters most for AWS revenue growth

5

. AWS has added more than 3.8 gigawatts in the past 12 months and plans to add over 12 gigawatts by year-end 2027, potentially supporting up to $150 billion in incremental annual AWS revenue if demand remains strong, according to Wells Fargo analysts

5

. Each incremental gigawatt translates to roughly $3 billion of annual cloud revenue

5

.

AWS re-accelerated to 20.2% year-over-year growth in Q3, the fastest pace since 2022, as the company addresses supply constraints that limited earlier expansion

5

. The dual approach of custom AI chips and on-premises AI infrastructure aims to maintain AWS's lead over Microsoft Azure and Google Cloud during intense competition for AI adoption

5

.

Market Competition and Timing Challenges

AWS faces stiff competition in the on-premises AI market. Dell's AI Factory with Nvidia captured 3,000 customers and shipped $15.6 billion in AI servers year to date, while HPE's private AI cloud won over 300 new customers

4

. Forrester analysts note that AI spending faces scrutiny as customers demand clear returns on investment, with free cash flow tightening and warnings of a potential dot-com-style bubble

4

. Microsoft has also deployed its own Nvidia AI Factories for OpenAI workloads and outlined data centers addressing data sovereignty in local countries . The trend toward enterprise datacenters and hybrid clouds represents an ironic shift for major cloud providers, reminiscent of infrastructure strategies from 2009 .

Today's Top Stories

TheOutpost.ai

Your Daily Dose of Curated AI News

Don’t drown in AI news. We cut through the noise - filtering, ranking and summarizing the most important AI news, breakthroughs and research daily. Spend less time searching for the latest in AI and get straight to action.

© 2025 Triveous Technologies Private Limited
Instagram logo
LinkedIn logo