Micron develops stacked GDDR memory as cost-effective HBM alternative for AI inference

2 Sources

Share

Micron is developing a vertically stacked GDDR memory architecture positioned between traditional GDDR and high-bandwidth memory (HBM). The company aims to introduce early prototypes by 2027, targeting AI inference workloads that prioritize memory capacity and cost-efficiency over peak bandwidth. This innovation could reshape memory solutions for GPUs and AI accelerators.

News article

Micron Pursues Vertically Stacked GDDR Memory Architecture

Micron is developing a new vertically stacked GDDR memory architecture that positions itself as an HBM alternative, according to recent industry reports

1

2

. The memory manufacturer aims to introduce early prototypes by 2027, with initial GDDR stacking expected to feature around four layers

2

. This longer-term effort signals Micron's strategy to diversify memory solutions for GPUs and AI accelerators while addressing growing memory demands of AI without the premium costs associated with traditional HBM.

The concept centers on vertically stacking memory dies within a single package to increase memory capacity and bandwidth without requiring the complex silicon interposer designs that HBM demands

1

. This approach could deliver a more scalable and cost-effective alternative to HBM, particularly for applications that don't require HBM's full performance envelope. The stacked GDDR solution won't match HBM in raw performance, but it would offer higher capacities that complement modern AI inference workloads

2

.

Targeting AI Inference Workloads and Memory Solution for GPUs

The development appears driven by AI inference workloads, which benefit more from increased memory capacity and cost-efficiency than from peak bandwidth

1

. While HBM currently dominates high-end AI training hardware due to its extremely high bandwidth and efficiency, it comes with higher production costs and more demanding packaging requirements. GDDR remains widely used in GPUs and certain AI inference accelerators thanks to its lower cost and simpler integration

1

.

By stacking memory dies, Micron could significantly boost per-package density while maintaining a relatively accessible price point compared to HBM-based solutions

1

. If successful, stacked GDDR memory may become a viable option for mid-range AI accelerators and future GPU designs, offering flexibility between conventional GDDR and premium HBM.

Technical Challenges and Memory Innovation Strategy

Micron faces technical hurdles in developing this DRAM innovation, particularly around thermal management and signal integrity if the company relies on wire bonding

2

. Stacking GDDR proves more challenging than stacking power-efficient modules like LPDDR5X, which Micron has already stacked up to 16-Hi in its SODIMM2 solution, achieving up to 256 GB per module

2

. The company may need to trade off clock speeds to address these technical constraints.

This development could intensify competition among memory manufacturers as the industry explores hybrid solutions to balance performance, cost, and scalability

1

. The timing is notable given that Micron experienced delays with HBM4 following certification issues with NVIDIA, while competitors like Samsung have seen increased supply allocation

2

. A GDDR-stacked solution could stand out in the memory market if it proves cost-effective relative to HBM, offering Micron a strategic differentiation point in the evolving landscape of high-performance computing and AI accelerators.

Today's Top Stories

TheOutpost.ai

Your Daily Dose of Curated AI News

Don’t drown in AI news. We cut through the noise - filtering, ranking and summarizing the most important AI news, breakthroughs and research daily. Spend less time searching for the latest in AI and get straight to action.

© 2026 Triveous Technologies Private Limited
Instagram logo
LinkedIn logo