AWS unveils AI Factories and Trainium3 chip while deepening Nvidia partnership for cloud dominance

Reviewed byNidhi Govil

18 Sources

Share

Amazon Web Services announced AI Factories for on-premises deployment and launched its Trainium3 chip at re:Invent 2025. The cloud giant revealed Trainium4 will support Nvidia's NVLink Fusion interconnect technology. These moves address data sovereignty concerns while AWS races to expand cloud capacity, adding over 12 gigawatts by 2027 to maintain its lead against Microsoft Azure and Google Cloud.

AWS Launches AI Factories to Address Data Sovereignty Concerns

Amazon Web Services unveiled AWS AI Factories at re:Invent 2025, a fully managed solution that brings AI infrastructure directly into corporate and government data centers

1

.

Source: SiliconANGLE

Source: SiliconANGLE

The service addresses data sovereignty requirements by allowing organizations to maintain complete control over their data while AWS handles hardware installation and management

3

. Customers provide the physical space and power, while AWS deploys its hardware and software, creating what functions as a private AWS Region on-premises

3

.

The on-premises AI solution emerged from AWS's work with Saudi Arabia's AI Zone, which will feature up to 150,000 AI chips and dedicated infrastructure

3

. AWS AI Factories combine Amazon's technology with Nvidia hardware, offering customers access to Blackwell GPUs or Amazon's Trainium3 chip, alongside AWS networking, storage, and services like Amazon Bedrock and SageMaker

1

. This approach mirrors similar offerings from Microsoft Azure, which has been deploying AI Factories in its global data centers for OpenAI workloads

1

.

Trainium3 Delivers Major Performance Gains for AI Workloads

AWS formally launched its Trainium3 UltraServer, powered by 3-nanometer AI chips that deliver significant improvements over previous generations

2

.

Source: The Register

Source: The Register

The AI training chip offers four times more compute performance, memory, and energy efficiency compared to Trainium2

2

. Each UltraServer hosts 144 chips, and thousands can be linked to provide up to 1 million Trainium3 chips for a single application—10 times the previous generation's capacity

2

.

Early customers testing Trainium3 have reduced AI training and inference costs by up to 50%

4

. Companies like Anthropic, Japan's Karakuri, Splashmusic, and Decart are already using the third-generation chip to significantly cut their inference expenses

2

. Each chip features 144 GB of HBM3E memory with 4.9 TB/s of memory bandwidth, capable of delivering over 2.5 petaFLOPS of dense FP8 performance

5

. The 40% improvement in energy efficiency addresses growing concerns about data center power consumption as AI workloads expand

2

.

Trainium4 Will Integrate Nvidia's NVLink Fusion Technology

AWS teased Trainium4, currently in development, which will support NVLink Fusion interconnect technology for seamless communication with Nvidia GPUs

2

.

Source: Market Screener

Source: Market Screener

This integration marks a significant deepening of the AWS and Nvidia partnership, allowing Trainium4-powered systems to interoperate with Nvidia hardware while using Amazon's lower-cost server rack technology

2

. The move could make it easier to attract AI applications built with Nvidia's CUDA platform, which has become the de facto standard

2

.

Amazon claims Trainium4 will deliver 3x more FLOPS at FP8, 6x the performance at FP4, and 4x the memory bandwidth

5

. The accelerators will work with Graviton CPUs and EFA networking technology across Nvidia's MGX racks

5

. While AWS hasn't announced a timeline, following previous patterns suggests more details will emerge at next year's re:Invent conference

2

.

Cloud Computing Capacity Expansion Drives AWS Strategy

Beyond new AI chips, AWS's primary focus remains rapidly expanding cloud computing capacity to maintain its lead against Microsoft Azure and Google Cloud

4

. AWS accelerated to 20.2% year-over-year growth in Q3, adding more than 3.8 gigawatts in the past 12 months

4

. Wells Fargo analysts estimate AWS will add more than 12 gigawatts of compute by year-end 2027, potentially supporting up to $150 billion in incremental annual revenue if demand remains strong

4

.

Oppenheimer analysts noted that each incremental gigawatt of compute added in recent quarters translated to roughly $3 billion of annual cloud revenue

4

. The capacity expansion could translate to 14% upside to 2026 AWS revenue and 22% upside in 2027

4

. However, AWS faces stiff competition in the AI infrastructure market, with Dell's AI Factory boasting 3,000 customers and $15.6 billion in AI server shipments year to date, while HPE gained over 300 new customers for its private AI cloud product

3

.

Forrester analysts warn that AI spending faces a reckoning as revenue from AI investments lags and customers demand clearer returns on investments that can cost millions to deploy and operate

3

. Despite these headwinds, AWS's dual strategy of custom AI chips and deepening Nvidia integration positions the company to serve diverse customer needs while managing capacity constraints that have limited growth.

Today's Top Stories

TheOutpost.ai

Your Daily Dose of Curated AI News

Don’t drown in AI news. We cut through the noise - filtering, ranking and summarizing the most important AI news, breakthroughs and research daily. Spend less time searching for the latest in AI and get straight to action.

© 2025 Triveous Technologies Private Limited
Instagram logo
LinkedIn logo