4 Sources
4 Sources
[1]
SK hynix completes development of HBM4 -- 2,048-bit interface and 10 GT/s speeds promised
SK hynix has completed development of its HBM4 memory and prepared it for high-volume manufacturing of memory stacks, the company said on Friday. HBM4 stacks from SK hynix go beyond specifications set by JEDEC by 25% in terms of performance, though it remains to be seen whether producers of next-generation AI accelerators -- such as AMD, Broadcom, or Nvidia -- will use that potential on their 2026 products. SK hynix's HBM4 memory stacks feature a 2,048-bit I/O, doubling the width of the HBM interface for the first time since 2015, and a 10 GT/s data transfer rate, which is 25% higher compared to the official JEDEC standard (more on this later). The company's HBM4 memory stacks use specially designed DRAM dies built using proven 1b-nm (5th Generation 10nm-class) process technology that combines decent performance efficiency with node maturity (i.e., low defect density and variability). For its HBM4 modules, SK hynix continues to use its proven Advanced Mass Reflow Molded Underfill (MR-MUF) method. In this process, multiple memory chips are placed on a base substrate and bonded together in a single reflow step. Right after that, spaces between the stacked DRAM layers, base die, and substrate are filled with a mold material to secure and protect the structure. Advanced MR-MUF enables the company to keep the height of its 12-Hi HBM stacks within specification and improve heat dissipation of power-hungry memory modules. "Completion of HBM4 development will be a new milestone for the industry," Joohwan Cho, Head of HBM Development at SK hynix, who has led the development, said. "By supplying the product that meets customer needs in performance, power efficiency and reliability in timely manner, the company will fulfill time to market and maintain competitive position." Unfortunately, SK hynix does not disclose either how many DRAM layers its HBM4 devices use, or their capacity, though we are likely talking about 12-Hi 36 GB devices that will be used for Nvidia's Rubin data center GPUs, but we are speculating. The company also does not disclose whether its initial HBM4 stacks use 12FFC+ (12nm-class) or N5 (5nm-class) base logic dies produced by TSMC. Arguably, the most interesting part of SK hynix's announcement is the fact that it decided to go beyond JEDEC's specification and certify its HBM4 stacks for a 10 GT/s data transfer rate, up from the 8 GT/s set by the standard. The company is not alone in going beyond JEDEC spec with HBM4: Micron is currently sampling HBM4 devices with a 9.2 GT/s data transfer rate, whereas Rambus has an HBM4 memory controller capable of 10 GT/s speed. While extra performance may be used by actual chip developers at some point, most developers apparently want to have some reserve for added piece of mind, according to an insight shared by Rambus a few years ago. SK hynix says that it is ready to mass produce its HBM4 stacks, though it does not disclose when it plans to kick off volume production of HBM4. "We are unveiling the establishment of the world's first mass production system of HBM4," said Justin Kim, President & Head of AI Infra at SK hynix. "HBM4, a symbolic turning point beyond the AI infrastructure limitations, will be a core product for overcoming technological challenges. We will grow into a full-stack AI memory provider by supplying memory products with the best quality and diverse performance required for the AI era in a timely manner."
[2]
SK Hynix says its HBM4 is ready for mass production
Top AI chipmakers count on faster, denser, more efficient memory to boost training AMD and Nvidia have already announced their next-gen datacenter GPUs will make the leap to HBM4, and if SK Hynix has its way, it'll be the one supplying the bulk of it. On Friday, the South Korean memory giant announced that it had wrapped HBM4 development and was preparing to begin producing the chips in high volumes. The news sent SK's share price on a 7 percent rally and for good reason. High Bandwidth Memory (HBM) has become an essential component in high-end AI accelerators from the likes of Nvidia, AMD, and others. Both Nvidia's Rubin and AMD's Instinct MI400 families of GPUs, pre-announced earlier this year, rely on memory vendors having a ready supply of HBM4 in time for their debut in 2026. The transition comes as the GPU slingers run up against the limits of existing HBM technologies, which currently top out at around 36 GB of capacity and about 1 TB/s of bandwidth per module, giving chips like Nvidia's B300 or AMD's MI355X about 8 TB/s of aggregate memory bandwidth. With the move to HBM4, we'll see bandwidth jump considerably. At GTC in March, Nvidia revealed its Rubin GPUs would pack 288 GB of HBM4 and achieve 13 TB/s of aggregate bandwidth. AMD aims to cram even greater quantities of memory onto its upcoming MI400-series GPUs, which will power its first rack-scale system called Helios. As we learned at AMD's Advancing AI event in June, the parts will pack up to 432 GB of HBM with an aggregate bandwidth approaching 20 TB/s. SK Hynix says that it has effectively doubled the bandwidth of its HBM by increasing the number of I/O terminals to 2,048, twice what we saw on HBM3e. This, it argues, has also boosted energy efficiency by more than 40 percent. While the DRAM typically in servers isn't usually a major energy consumer, HBM is. With the shift from 24 GB on AMD's MI300X to the 36 GB modules found on the MI325, power consumption jumped from 250 W to roughly a kilowatt per GPU. SK Hynix says that, along with more I/O terminals and improved efficiency, its chips have also managed to exceed the JEDEC standard for HBM4, achieving 10 Gb/s operating speed. Which of the three big HBM vendors will end up supplying these chips remains to be seen. While SK Hynix has won the majority of Nvidia's HBM business over the past few years, Samsung and Micron are also working to bring HBM4 to the market. Micron began sampling 36 GB 12-high HBM4 stacks to customers in June. Much like with SK Hynix, the stacks are using a 2048-bit interface and will achieve roughly twice the bandwidth of the HBM3e modules available today. The American memory vendor expects to ramp production of the stacks sometime next year. Meanwhile, for Samsung, HBM4 presents a new opportunity to win over Nvidia's business. The vendor has reportedly struggled to get its HBM3e stacks validated for use in the GPU giant's Blackwell accelerators. ®
[3]
SK Hynix says readying HBM4 production after completing internal certification
SEOUL, Sept 12 (Reuters) - South Korea's SK Hynix (000660.KS), opens new tab said on Friday it has completed its internal certification process for next generation high-bandwidth memory 4(HBM4) chips and established a production system for customers. In March, the South Korean chipmaker, a key supplier to AI giant Nvidia (NVDA.O), opens new tab, said it had shipped its HBM4 12-layer chip samples to customers, adding that the company aims to complete preparations for mass production of 12-layer HBM4 products within the second half of this year. HBM - a type of dynamic random access memory or DRAM standard first produced in 2013 - involves stacking chips vertically to save space and reduce power consumption, helping to process the large volumes of data generated by complex AI applications. SK Hynix shares hit a record high in morning trade, rising as much as 7.3%, outpacing the benchmark KOSPI's 1.2% rise, as analysts cited the company's HBM4 chip production plans. Kim Sunwoo, a senior analyst at Meritz Securities, projected that SK Hynix's HBM market share will remain in the low 60% range in 2026, supported by early HBM4 supply to key customers and the resulting first-mover advantage. SK Hynix is currently the main HBM supplier to Nvidia, although Samsung Electronics (005930.KS), opens new tab and Micron (MU.O), opens new tab supply it with smaller volumes. An SK Hynix executive told Reuters in an interview last month that because of technological changes in how SK Hynix and rivals such as Micron and Samsung build next-generation HBM4, their products include a customer-specific logic die, or "base die", that helps manage the memory. That means it is no longer possible to easily replace a rival's memory product with a nearly identical chip or product. Samsung said in July that it had provided samples of its HBM4 chips to customers, with a plan to supply them next year. Reporting by Heekyong Yang and Jihoon Lee; Editing by Ed Davies Our Standards: The Thomson Reuters Trust Principles., opens new tab
[4]
SK hynix Inc. Completes First HBM4 Development and Readies Mass Production
SK hynix Inc. announced that it has completed development and finished preparation of HBM4, a next generation memory product for ultra-high performance AI, mass production for the world's first time. HBM(High Bandwidth Memory): This high-value, high-performance memory vertically interconnects multiple DRAM chips and dramatically increases data processing speed in comparison to conventional DRAM products. There are six generations of HBM, starting with the original HBM which was followed by HBM2, HBM2E, HBM3, HBM3E, and HBM4. SK hynix said that the company has successfully completed development and based on this technological achievement, the company has prepared HBM4 mass production to lead the AI era. Through this momentum, the company has once again proven the AI memory leadership in the global market. With recent dramatic increase in AI demand and data processing, the needs for high bandwidth memory for faster system speed are surging. In addition, securing memory power efficiency has emerged as a key requirement for customers as power consumption for data center operation has increased. SK hynix expects HBM4, with increased bandwidth and power efficiency, to be the optimal solution to meet customers' needs. Bandwidth: In HBM products, bandwidth refers to the total data capacity that one HBM package can process per second. HBM4, of which mass production system has been readied, has the industry's best data processing speed and power efficiency with the bandwidth doubled through adoption of 2,048 I/O terminals, double from the previous generation, and power efficiency improved by more than 40%. The company expects to improve AI service performance by up to 69% when the product is applied, which will lead to solve data bottleneck and significantly reduce data center power costs. The company has far exceeded the JEDEC standard operating speed(8Gbps) by implementing over 10Gbps(Gigabit per second) operating speed in HBM4. JEDEC(Joint Electron Device Engineering Council): A standardization body that is the global leader in developing open standards and publications for the microelectronics industry. In addition, the company adopted the Advanced MR-MUF process in HBM4, which has been proven to be reliable in the market, and the 1bnm process, or the fifth-generation of the 10-nanometer technology, to minimize the risk in mass production. MR-MUF(Mass Reflow Molded Underfill): The process of stacking semiconductor chips, injecting liquid protective materials between them to protect the circuit between chips, and hardening them. The process has proved to be more efficient and effective for heat dissipation, compared with the method of laying film-type materials for each chip stack. SK hynix's advanced MR-MUF technology is critical to securing a stable HBM mass production as it provides good warpage control and reduces the pressure on the chips being stacked.
Share
Share
Copy Link
SK hynix announces completion of HBM4 development and readiness for mass production, promising significant improvements in bandwidth and efficiency for AI applications.
SK hynix, the South Korean memory giant, has announced a significant breakthrough in the development of High Bandwidth Memory (HBM) technology. The company has successfully completed the development of HBM4 and is now prepared for high-volume manufacturing of memory stacks
1
2
. This achievement marks a crucial step forward in meeting the growing demands of AI accelerators and data centers.Source: Tom's Hardware
The new HBM4 memory stacks from SK hynix boast impressive specifications that go beyond the JEDEC standard:
1
.1
4
.4
.These advancements are made possible by SK hynix's use of their 1b-nm (5th Generation 10nm-class) process technology for DRAM dies and the continued implementation of their proven Advanced Mass Reflow Molded Underfill (MR-MUF) method
1
4
.Source: The Register
The development of HBM4 is expected to have a significant impact on AI and data center performance:
4
.4
.2
.Related Stories
SK hynix's announcement has already had a positive impact on its market position:
2
3
.3
.3
.Source: Reuters
While SK hynix has not disclosed specific details about the number of DRAM layers or the capacity of its HBM4 devices, industry speculation suggests they may be 12-Hi 36 GB devices
1
. The company has stated that it has established a production system for customers and is ready for mass production3
4
.As the AI industry continues to evolve rapidly, the availability of HBM4 from SK hynix and its competitors will play a crucial role in enabling the next generation of high-performance AI accelerators and data center GPUs.
Summarized by
Navi
[1]
[2]
[4]
1
Business and Economy
2
Business and Economy
3
Technology