Curated by THEOUTPOST
On Fri, 11 Oct, 12:05 AM UTC
16 Sources
[1]
AMD's AI networking solutions aim to enhance performance and scalability in AI environments - SiliconANGLE
AMD's AI networking solutions aim to enhance performance and scalability in AI environments Advanced Micro Devices Inc. last week introduced its next-generation AI networking components -- the Pensando Salina data processing unit and the Pensando Pollara 400 AI network interface card, both of which are core components in the artificial intelligence infrastructure landscape. The components address the growing demands of hyperscale data centers and AI workloads, with implications for the broader market and competitors. I see four impacts on the market from AMD's introduction, including: AMD's announcement will send ripples through the industry as a whole, with competitive pressures and consolidation likely outcomes. The introduction of the AMD Pensando Salina DPU and Pollara 400 AI NIC reflects AMD's broader strategic efforts to carve out a more significant share of the AI infrastructure market. With AI workloads expanding in complexity and scale -- particularly in hyperscale data centers -- the ability to optimize data flow both at the front end (to AI clusters) and the back end (accelerator-to-accelerator communication) will be crucial for maintaining performance and efficiency. AMD's next-gen networking solutions represent a step toward more efficient and scalable AI infrastructures. Still, their long-term market position will depend on how they perform in real-world deployments and whether AMD can maintain a pace of innovation that keeps it ahead of the competition. In an industry driven by rapid technological advancements, this balance of performance, scalability and adaptability will determine whether AMD can solidify its place in the AI networking business. From a competitive standpoint, though AMD's solutions check the performance box, the moat around Nvidia has more to do with software and the ecosystem than with who has the fastest chips. AMD is on the right path, but putting a dent in Nvidia is a long road to hoe. Intel is another story. Though some industry watchers are bullish on the outlook for the former market leader, I am reminded of the Michael Scott quote, "Fool me once, strike one, fool me twice, strike three." My prediction is that the AI arms race will be fueled by Nvidia and AMD with Intel on the outside looking in unless something dramatically changes.
[2]
AMD Looks to Challenge Nvidia With a New Chip. Can the Stock Be the Next Big AI Chip Winner? | The Motley Fool
Here's why AMD is well positioned to benefit from the AI infrastructure buildout. Advanced Micro Devices (AMD -1.56%) recently introduced its latest artificial intelligence (AI) chip as it looks to make a dent in Nvidia's (NASDAQ: NVDA) huge market share. While AMD, as it is commonly called, has seen its AI-related revenue soar this year, it's still just a fraction of the revenue generated by its rival's graphic processing units (GPUs). The question is, can this new chip help AMD become the next big AI winner? AMD has steadily increased its revenue for data center GPUs this year from an original forecast of more than $3.5 billion to most recently $4.5 billion. However, that pales in comparison to the $26.3 billion in data center revenue that Nvidia generated just last quarter. The semiconductor company is hoping that its newest chip, called the Instinct MI325X, will help close some of the large share gap in GPUs between the two companies. AMD said the new chip offers industry-leading memory capacity and bandwidth and more compute power than Nvidia's H200 chip. However, it is notable that the comparison is to Nvidia's Hopper architecture and not its newest Blackwell architecture. For its part, Nvidia claims Blackwell can have 30 times the inference performance as Hopper. AMD said the new chip will begin production in Q4 and be ready for widespread shipment in early 2025. It said the chips will be available through a wide set of platform providers, including Dell, Super Micro Computer, and Hewlett Packard Enterprise, among others. Like Nvidia, AMD is moving its AI chip development cycle up to about once a year. The company also introduced some networking solutions, including a new NIC (network interface card) and DPU (data processing unit). NICs are needed to connect to a network, while DPUs handle data transfers, compression, and storage. It also introduced a new line of CPUs (central processing units) called EPYC 5th Gen, which it said has been designed to accelerate data center, cloud, and AI workloads. Meanwhile, AMD said it is adding some powerful new features and capabilities to its ROCm open software stack. While AMD has produced powerful GPUs in the past, Nvidia's CUDA software has long been a huge advantage for the company, since developers have long been trained to program GPUs through its platform. This has created a wide moat for Nvidia that AMD and others have struggled to compete against. AMD did not announce any new customers for its new products, but representatives from Alphabet, Meta Platforms, Microsoft, and OpenAI did discuss how they are using AMD at the event. Meta noted that it is "working on several training workloads with AMD." There was nothing from AMD's announcement that indicates it will take any meaningful share from Nvidia in the near future. However, the AI infrastructure market is expanding so quickly and Nvidia only has so much supply capacity, that there will still be a nice opportunity for AMD to grow its AI chip business. In addition, companies will want to keep Nvidia honest and have at least a second GPU supplier, so that it does not become a complete monopoly. There were also some comments from customers that suggest AMD's GPUs are being given some more important workloads than in the past. This combined with the strong overall market for GPUs should see the company continue to produce solid AI revenue growth, just off a much lower base than Nvidia. And at the end of the day, AMD doesn't need to take a ton of share from Nvidia to be an AI winner. Going from a 4% market share to 7% in a widely expanding market would still be a lot of growth. From a valuation standpoint, AMD currently trades at a forward price-to-earnings (P/E) ratio of about 30 based on 2025 analyst estimates. That's a fair valuation given the potential growth the company has in front of it from AI. AMD is likely not going to be the next Nvidia, but it is a solid secondary option in the space for investors who want to diversify their AI stock holdings. As long as the market for AI infrastructure is growing strongly, the stock should be a solid performer.
[3]
AMD Stock: Can Its AI Strategy Compete With Nvidia? Analyst Predicts Market Share Gains Ahead - NVIDIA (NASDAQ:NVDA), Advanced Micro Devices (NASDAQ:AMD)
AMD's new MI325X GPU targets Nvidia's AI dominance, delivering 1.4x better inferencing performance than Nvidia's H200. Advanced Micro Devices Inc AMD just unveiled its next wave of AI weaponry, and it's taking aim at none other than Nvidia Corp NVDA. AMD is gearing up to claim a bigger slice of the AI pie in an AI event packed with hardware reveals, partnerships, and tantalizing hints of future dominance. But can the underdog challenge the reigning champ in AI accelerators? JPMorgan's Harlan Sur remains cautious, sticking with a Neutral rating while highlighting the hefty potential in AMD's AI-focused arsenal. AI Market Outlook: Bigger, Better, Faster AMD's AI event wasn't just a product showcase; it was a declaration of intent. The company bumped up its AI accelerator total addressable market (TAM) outlook from $400 billion in 2027 to a massive $500 billion by 2028, thanks to a continued surge in AI model deployments and use cases, Sur noted. The TAM growth may decelerate as we inch toward 2028, but AMD's strong foothold with its CPUs, GPUs, and DPUs puts it in a prime position to capture more of the AI wave. The recent ZT Systems acquisition will also help AMD iron out any kinks in production and speed up customer ramp-ups -- a clear sign the company is focused on scaling fast, the analyst said. Read Also: What Happened With AMD Stock Today? MI325X GPU: AMD's AI Champion Steps Into The Ring The event's highlight? AMD's MI325X, a GPU designed to go head-to-head with Nvidia's H200 HGX in AI training. While the MI325X holds its own in training, it takes the crown in AI inferencing, outpacing Nvidia's H200 by 1.4x. Production shipments start soon, and with the MI350 set for 2025 -- boasting a 1.8x boost in AI compute performance -- AMD is clearly not backing down in its GPU battle with Nvidia. But, as Sur points out, this race is a marathon, not a sprint. AMD is ramping up, but Nvidia's lead still looms large, and capturing market share will take time. Server CPUs: AMD's Zen-Fueled Takeover AMD isn't leaving its core CPU game behind, either. The company launched its next-gen EPYC processor, codenamed "Turin," built on its Zen 5 architecture. With a 40% performance edge over Intel's Granite Rapids, AMD continues to eat up server CPU market share. Sur noted that AMD's server CPU share jumped from 31% to 34% in the first half of 2024, and Turin could drive further gains in the coming quarters. Software, Strategic Partnerships: The Secret Sauce Beyond hardware, AMD's progress on the software front is crucial. The ROCm 6.2 software stack has shown a 2.4x improvement in AI inference performance -- critical for staying competitive in a market where software optimization can be a game-changer. Add in deep partnerships with cloud giants like Meta Platforms Inc META, Microsoft Corp MSFT, and Oracle Corp ORCL, and AMD has the strategic backing to take on the big leagues. The Bottom Line? AMD's AI event showcased a tech company ready for battle, but dethroning Nvidia will not be easy. Sur's price target of $180 by the end of 2025 reflects optimism but caution as AMD's ramp-up in AI accelerators and servers plays out over the next few years. Keep an eye on AMD stock as it gears up for its biggest AI push yet -- this might be a battle worth watching. Read Next: Nvidia's Blackwell Chip Faces AMD's MI350 Challenge In 2025: CEO Lisa Su Says, 'Beginning, Not The End Of The AI Race' Photo: Shutterstock Market News and Data brought to you by Benzinga APIs
[4]
AMD attacks AI workloads with next-generation Instinct MI325X accelerator and networking chips - SiliconANGLE
AMD attacks AI workloads with next-generation Instinct MI325X accelerator and networking chips As demand for the powerful hardware needed to support artificial intelligence workloads continues to surge, Advanced Micro Devices Inc. is stepping up its game with the launch of its most sophisticated AI accelerator yet. At a media and analyst event called "Advancing AI" in San Francisco today, the company showed off its next-generation AI chip, the AMD Instinct MI325X accelerator, alongside a new networking platform based on the AMD Pensando Salina data processing unit. The new technologies are set to launch next year, with AMD promising that they'll set a new standard in terms of generative AI performance. They're part of AMD's continued push into data center chips -- an area where it has stolen a march on longtime rival Intel Corp., capturing 34% of data center chip revenues -- by focusing on making them better for AI. "Our goal is to make AMD the end-to-end AI leader," Chief Executive Lisa Su (pictured, holding an AMD EPYC chip), said this morning at her event keynote. She trotted a Who's Who of executives at AI leaders such as Microsoft Corp., Meta Platforms Inc., Databricks Inc. and Oracle Corp., as well as startups such a Reka AI Inc., Essential AI Labs Inc., Fireworks AI and Luma AI Inc., who lauded their partnerships with AMD. The chipmaker, which has rapidly emerged as a growing threat to the dominance of Nvidia Corp. in the AI infrastructure industry, is building on the success of its fast-selling MI300X AI chip, which launched earlier this year and is forecast to drive more than $4 billion in AI chip sales for the company. The AMD Instinct MI325X, built on the company's CDNA 3 architecture, is designed to enable blazing-fast performance combined with greater energy efficiency for the most demanding AI tasks, including training large language models, fine-tuning their performance and AI inference, where the models deliver results to users. In a press briefing, Brad McCredie, AMD's corporate vice president of GPU platforms, said the MI325X chip brings more than double the performance of the existing MI300 in terms of performance for AI inference and training workloads. McCredie proceeded to reel off a list of impressive numbers, saying the MI325X features 256 gigabytes of high-performance HBM3E memory and up to 6 terabits per second in terms of bandwidth. That equates to 1.8 times more capacity and 1.3 times more bandwidth than Nvidia's current most powerful AI chip, the Nvidia H200 graphics processing unit. In addition, the MI325X provides 1.3 times greater peak theoretical FP16 and FP8 performance than its most powerful competitor. "This industry-leading memory and compute can deliver up to 1.3 times the inference performance of the H200 on Mistral 7B at FP16, and 1.2-times inference performance on Llama 3.1 70B at FP8," McCredie said. Those numbers suggest the AMD MI325X chips will pack a powerful punch for AI developers, and the best news is they won't have to wait that long to see how well they stack up. AMD said the MI325X is slated to enter production by the fourth quarter, with companies such as Dell Technologies Inc., Hewlett Packard Enterprise Co., Lenovo Group Ltd., Gigabyte Technology Co. Ltd. and Super Micro Computer Inc. planning to put servers using them on sale by the first quarter of next year. How much they will challenge Nvidia, however, remains to be seen. "AMD certainly remains well-positioned in the data center, but I think their CPU efforts are still their best-positioned products," said Ben Bajarin, CEO and principal analyst at Creative Strategies Inc. "The market for AI acceleration and GPUs is still heavily favoring Nvidia and I don't see that changing anytime soon." AMD intends to pair the latest Instinct accelerators with new networking technologies, including the AMD Pensando Salina DPU and the AMD Pensando Pollara 400, which it claimed is the industry's first Ultra Ethernet-Ready AI network interface card. The new technologies are critical for linking AMD's new AI accelerators and ensuring sufficient throughput for sharing data. The AMD Pensando Salina DPU represents the front end of AMD's network, which delivers data to clusters of Instinct accelerators, while the AMD Pensando Pollara 400 represents the back end, which manages data transfer between individual accelerators and clusters. They'll be available early next year. The Pensando Salini DPU is the third-generation of its DPU series, delivering twice the performance of its predecessor and more than doubling the available bandwidth and scale. All told, it supports 400G throughput, ensuring some of the fastest data transfer rates ever seen in a data center. As such, it will serve as a critical component of AI front-end network clusters, helping to optimize the performance, efficiency and scalability of AI applications. In the briefing, Soni Jiandani, senior vice president and general manager of AMD's Network Technology and Solutions Group and a co-founder of Pensando Systems Inc. that AMD acquired in 2022, stressed the importance of networking. She explained that AI systems need to connect to the front end of the network for users, while at the back end they must be linked to thousands of GPUs to ensure performance. "Back-end networks drive AI system performance," she said. "Meta says 30% of its training cycle time typically elapses while waiting for networking. So networking is not only critical, it's foundational to driving AI performance." IDC analyst Brandon Hoff agreed. "AI workloads, especially generative AI workloads, are the first to be able to consume all of the compute, memory, storage and networking in a server node," he explained. "AI also scales beyond a single AI Factory node which requires all of the GPUs to talk to each other." As a result, he added, "the time to communicate between AI Factory nodes is called 'time in network' and can be up to 60% of the processing time for a training or multi-node inferencing AI run. To put it a different way, if a hyperscaler spent $1 billion on their GPUs, they get $400 million of work done and $600 million of the GPUs sitting idle. High-performance networking is essential, and the second most important piece." So for AMD, he said, "having a strong networking set of products is an essential part of building their AI business. These are the right products for DPUs and SmartNICs, and Ethernet is the right technology to invest in." Bajarin said AMD is making good progress technically in networking. "I imagine the more AMD can integrate this into their full-stack approach to optimizing for the racks via the ZT systems purchase, then I think their networking stuff becomes even more important," he said. The Salina DPU and the Pollara 400 will also launch early next year, but while companies are waiting for those technologies to become available, they can at least contemplate what's coming further along down the line. In addition to its new chips and networking technologies coming out shortly, AMD also previewed the next of its next-generation chips, the AMD Instinct MI350 series accelerators, slated to arrive in the second half of next year. According to AMD, the Instinct MI350 will be built on the company's CDNA 4 architecture and deliver an incredible 35-times improvement in terms of inference performance, compared with the Instinct MI325X, while providing 288GB of HBM3E memory. The company also mentioned plans for a new MI355X accelerator chip that's expected to start shipping in volume in late 2025, followed by the MI400 chip that will be based on an entirely new architecture when it launches sometime in 2026. Though everyone's focused on AI, AMD has no intention to let up on its wider assault of the market for data center servers. At the event, it also lifted the lid on its latest EPYC central processing unit, formerly codenamed "Turin," aimed at enterprise, AI and cloud workloads. The 5th Gen AMD EPYC Series processors will be built on the company's Zen 5 core architecture, with the company aiming to offer a wide range of core counts to suit different use cases, spanning from as low as just eight all the way up to 192 cores for the most demanding workloads. The company promised that 5th Gen EPYC chips will build on the performance of its existing, 4th Gen EPYC platform, with its biggest, 192-core EPYC 9005 Series offering 2.7 times better performance than its most powerful existing chip. Meanwhile, the new 64-core EPYC 9575F is customized for GPU-powered AI workloads, offering 5 gigabits per second in terms of performance, providing 28% faster processing than competing chips. Dan McNamara, senior vice president and general manager of AMD's server business, said customers can put their trust in the company's performance claims. "With five generations of on-time roadmap execution, AMD has proven it can meet the needs of the data center market and give customers the standard for data center performance, efficiency, solutions and capabilities for cloud, enterprise and AI workloads," he said. Finally, AMD teased the imminent launch of its third-generation of mobile processors laptops and notebooks. The new Ryzen AI Pro 300 Series processors are built on an advanced four-nanometer process and are said to be powerful enough to support on-device AI workloads such as generative AI copilots, live captioning and AI-powered translations. They offer a threefold increase in AI performance from the previous-generation Ryzen chips, together with advanced security and manageability features for enterprise users. Like the latest EPYC chips, the new Ryzen processors are built on AMD's Zen 5 architecture, and they'll provide up to 40% better performance and up to 14% faster productivity performance compared with the Intel Core Ultra 7 165U. They'll also "significantly" extend device battery life, the company added. In addition, they'll come with an integrated neural processing unit that can deliver 50 trillions of operations per second in AI processing power, exceeding Microsoft Corp.'s requirements for its AI-powered Copilot tools, AMD said. Jack Huynh, senior vice president and general manager of AMD's Computing and Graphics Group, said the PRO 300 Series is designed to address enterprise's increased demands for more compute power and efficiency on their business machines. "Our third-generation AI-enabled processors for business PCs deliver unprecedented AI processing capabilities with incredible battery life and seamless compatibility for the applications users depend on," he said. The bottom line, said Creative Strategies' Bajarin, is that "the data center is under a complete transformation and we are still only in the early days of that which makes this still a wide open competitive field over the arc of time 10-plus years. I'm not sure we can say with any certainty how this shakes out over that time but the bottom line is there is a lot of market share and money to go around to keep AMD, Nvidia and Intel busy."
[5]
AMD Challenges Nvidia With New AI Chip | PYMNTS.com
Advanced Micro Devices (AMD) on Thursday (Oct. 10) unveiled a new artificial intelligence (AI) chip aimed at breaking Nvidia's stronghold on the lucrative data center GPU market. The launch of AMD's Instinct MI325X accelerator marks an escalation in the AI hardware arms race, with implications for businesses investing in AI. AMD's announcement came during its Advancing AI 2024 event, where the company revealed a broad portfolio of data center solutions for AI, enterprise, cloud and mixed workloads. This portfolio includes the new Instinct MI325X accelerators, 5th Gen AMD EPYC server CPUs, AMD Pensando Salina DPUs, AMD Pensando Pollara 400 NICs and AMD Ryzen AI PRO 300 series processors for enterprise AI PCs. The generative AI boom, fueled by technologies like large language models, has created a high demand for powerful GPUs capable of training and running complex AI systems. Nvidia has been the primary beneficiary of this trend, with its data center revenue jumping in recent earnings reports. "Nvidia's dominant position in the AI chip market has remained virtually unchallenged," Max (Chong) Li, an adjunct professor at Columbia University and founder CEO of decentralized AI data provider Oort, told PYMNTS. "AMD's new chip should at least provide some competition, which could lead to pricing pressure in the long term. Some reports have estimated that Nvidia earns as much as a 75% profit margin on AI chips. Should AMD start to eat market share, one would assume prices will begin to drop, as they often do across most industries when companies compete for customers." The CUDA ecosystem is Nvidia's proprietary parallel computing platform and programming model, which has become the standard for AI and high-performance computing tasks. AMD's challenge extends beyond hardware performance to providing a compelling software ecosystem for developers and data scientists. AMD has invested in its ROCm (Radeon Open Compute) software stack, reporting at the event that it has doubled AMD Instinct MI300X accelerator inferencing and training performance across popular AI models. The company said over one million models run seamlessly out of the box on AMD Instinct, triple the number available when MI300X launched. "AMD's launch of the Instinct MI325X chip marks a significant step in challenging NVIDIA's dominance in the data center GPU market, but it's unlikely to dramatically alter the competitive landscape immediately," Dev Nag, CEO of QueryPal, a support automation company, told PYMNTS. "NVIDIA's 95% market share in AI chips is deeply entrenched, largely due to their mature and dominant CUDA ecosystem. "The success of AMD's initiative hinges not just on the performance of their chips, but on their ability to address the software side of the equation," he added. "NVIDIA spends about 30% of its R&D budget on software and has more software engineers than hardware engineers, meaning that it will continue to push its ecosystem lead forward aggressively." AMD's entry into the market could affect businesses looking to adopt AI technologies. Increased competition might lead to more options and better pricing in the long term. Nag suggested that over the next 2-3 years, "as AMD refines its offerings and potentially gains market share, we could see more options at various price points. This could make AI hardware more accessible to small and medium-sized enterprises that have been priced out of the current market." According to Nag, immediate price drops are unlikely. "Current demand for AI chips far outstrips supply, giving manufacturers little incentive to lower prices," he told PYMNTS. "AMD appears to be positioning itself as a value option rather than significantly undercutting Nvidia on price." AMD's focus on open standards could have broader implications. "If successful, it could lead to more cost-effective solutions by reducing dependency on proprietary ecosystems like CUDA," Nag said. "This approach could encourage more interoperability and flexibility in AI development, potentially making it easier for businesses to adopt and integrate AI solutions." Industry partners have responded positively to AMD's announcement. The company showcased collaborations with major players, including Dell, Google Cloud, HPE, Lenovo, Meta, Microsoft, Oracle Cloud Infrastructure and Supermicro. AMD Chair and CEO Lisa Su said in a statement that the data center AI accelerator market could grow to $500 billion by 2028. Even a tiny slice of this market could represent significant revenue for AMD, making its push into AI chips a critical strategic move. For businesses across various sectors, from retail to manufacturing, developing a more competitive AI chip market could speed up the integration of AI into core operations and customer-facing services. More accessible and powerful AI hardware could make tasks like demand forecasting, process optimization and personalized customer experiences more feasible for a broader range of companies. "Lower prices always lower barriers to entry and enable more businesses and people to take advantage of newer technologies," Li said. "Take, for example, mobile phones. Back when they first debuted, the public's view of a mobile phone user was that of a wealthy person in a fancy car making calls on the go. Now, most people in developed and many people in emerging countries tend to have at least a basic smartphone; soon, access to AI is likely to experience a similar adoption boom."
[6]
AMD Launches New Artificial Intelligence (AI) Chips: Should Nvidia Investors Be Worried? | The Motley Fool
AMD is trying to make a dent in the AI graphics card market, but will it be able to eat into Nvidia's market share? The market for artificial intelligence (AI) accelerators is currently dominated by Nvidia (NVDA -4.69%), which has built a technology lead over rivals and controls an estimated 85%-plus share of this space. However, fellow chipmakers, such as Advanced Micro Devices (AMD -5.22%) and Intel, are also trying to corner a piece of this lucrative market. AMD recently unveiled a new AI chip -- the Instinct MI325X -- claiming that it's better than Nvidia's current flagship H200 processor. However, shares of AMD sank following the announcement. Read on to discover why that may have been the case and check if AMD's latest AI chip should be a cause for concern for Nvidia investors. AMD points out that its MI325X data center accelerator is equipped with 256 gigabytes (GB) of HBM3E high-bandwidth memory that supports a bandwidth of 6 terabytes (TB) per second. According to AMD, its latest chip has 1.8x more capacity and 1.3x more bandwidth than Nvidia's flagship H200 processor, which is based on the Hopper architecture. As a result, AMD claims that the MI325X offers 1.3x greater compute performance, theoretically, along with an identical advantage in AI inference on the Llama 3.1 and Mistral 7B large language models (LLMs). The company is on track to start the production of the MI325X from the current quarter and says it will be widely available in the first quarter of 2025. Given that Nvidia's H200 processor offers 141GB of HBM3E memory at a bandwidth of 4.8TB/second, there's no doubt that AMD's new AI chip is superior on paper. But at the same time, investors should note that the H200 was announced nearly a year ago and started shipping in the second quarter of 2024. AMD's chip, therefore, is around nine months later to the market than the H200. On the other hand, Nvidia is set to level up its game in the AI chip market with its next-generation Blackwell processors. The upcoming B200 Blackwell graphics processing unit (GPU) is set to pack 208 billion transistors, as compared to the 153 billion transistors on the MI325X. More importantly, the B200 is reportedly manufactured using Taiwan Semiconductor Manufacturing's 4-nanometer (nm) 4NP process node, compared to the 5-nm process node for the MI325X. Nvidia will be packing a much higher transistor count into a much smaller chip, which means that its B200 processor is likely to have more computing power and be more power-efficient at the same time. Also, the B200 is reportedly going to offer a higher memory bandwidth of 8TB/second. The good part is that Nvidia expects to "get several billion dollars in Blackwell revenue" in the fourth quarter of its ongoing fiscal year, which will run from November 2024 to January 2025. What this means is that Nvidia would have moved ahead of AMD by one generation by the time the latter's latest chip is widely available to purchase. As a result, there doesn't seem to be a strong reason for Nvidia investors to be wary of AMD's latest offerings. However, there does appear to be a silver lining for AMD investors. Nvidia is running away with the AI chip market and seems on course to clock almost $100 billion in data center revenue in the current fiscal year. AMD, on the other hand, is expecting its 2024 AI data center GPU revenue to hit $4.5 billion, which pales in comparison to Nvidia's potential revenue from this segment. However, AMD doesn't need to beat Nvidia in AI GPUs to supercharge its growth. It simply needs to become the second-biggest player in this market, which AMD believes could be worth a whopping $500 billion in 2028. The company started selling its AI GPUs toward the end of 2023 and sold $6.5 billion worth of data center chips last year, including its Epyc server processors. This year, the company has generated $5.2 billion in data center revenue already in the first six months of 2024, which is nearly double its data center revenue in the same period last year. At this run rate, AMD could generate nearly $10.5 billion in data center revenue this year, $4.5 billion of which will come from sales of AI GPUs. Assuming AMD could capture even 10% of the AI GPU market in 2028, it could generate $50 billion in annual revenue from this segment, which would be a big jump from this year's projections. Additionally, AMD's MI325X processor could witness robust demand, as Nvidia pointed out on its previous earnings conference call that its H200 processors are expected to keep selling nicely despite the arrival of the new Blackwell chips. More specifically, Nvidia estimates that shipments of Hopper-based chips, such as the H200, are set to rise in the second half of the current fiscal year, suggesting that there's space for the MI325X, despite the arrival of more powerful processors. Moreover, AMD's growing presence in the AI GPU market is expected to boost its growth from next year. While the company's revenue is projected to jump 13% in 2024 to $25.6 billion, the forecast for the next couple of years appears to be even better. Even if AMD is unable to beat Nvidia in the AI chip market, it could turn out to be a solid AI stock in the long run if it manages to carve out a small niche for itself. At the same time, investors should note that AMD can capitalize on other AI-related applications, as well in the form of AI-enabled PCs and server processors. Even though AMD stock has delivered tepid gains of just 14% in 2024, investors would do well to keep it on their watch lists as it could soar impressively, thanks to the huge opportunity in AI GPUs and other markets.
[7]
AMD says new AI chips will be out soon and can outperform Nvidia
Advanced Micro Devices, looking to catch up with Nvidia in the lucrative market for artificial intelligence processors, said its latest chips are rolling out to data centers and will exceed some of the capabilities of its rival. Computer systems based on AMD's MI325X processors will be available soon and have an edge over machines running Nvidia's H100, CEO Lisa Su said at a company event in San Francisco on Thursday. The MI325X's use of a new type of memory chip will give it better performance at running AI software -- a process known as inference -- she said. The Santa Clara, California-based company is trying to crack Nvidia's dominance in so-called AI accelerators -- chips that have become essential to the development of artificial intelligence systems. Like Nvidia, AMD has committed to bringing out new accelerators every year, stepping up its pace of innovation. Still, AMD has a long way to go to match Nvidia. Under Su, who just marked her 10th anniversary in the top job at AMD, the company has eclipsed its longtime nemesis Intel in market valuation. But both companies were caught off guard by how ferociously the industry embraced AI accelerators. Of the two, AMD has responded far more quickly and established itself as the closest rival to Nvidia. AMD, which is expected to report quarterly results in the coming weeks, has set a target of $4.5 billion of revenue from the new type of chips for this year, a rapid increase. Su has said the overall market for such chips will hit $400 billion in 2027. On Thursday, she said the company expects that number to reach $500 billion in 2028. At the event, Su also said the company is releasing a new line of server processors based on its "Turin" technology, making a fresh push into a market once dominated by Intel. Computers are going on sale with AMD's fifth-generation EPYC central processing units, or CPUs, she said. The chips have as many as 192 processor cores and can outperform the latest Intel products, she said. The company said it has 34% of the market for this category of chips when measured by revenue. Though Intel still dominates the segment, it once had a 99% share.
[8]
AMD Gains Momentum With AI: Can It Beat Expectations? - Advanced Micro Devices (NASDAQ:AMD), Hewlett Packard (NYSE:HPE)
Advanced Micro Devices AMD is gaining traction with AI. NVIDIA remains the leader because of its first-mover advantage, but there are other significant players, and AMD is the leader. While NVIDIA commands an estimated 95% of the AI data center market, AMD carries most of the remainder and can regain the share lost due to NVIDIA's boom. That's worth more than 1000 basis points in market share growth on top of robust organic industry growth to drive revenue, earnings, and shareholder value. Industry growth is still accelerating. AMD CEO Lisu Su upped her estimates at the investor day event, expecting a 60% CAGR over the next five years. Aside from company commentary, there is solid evidence of the demand for AMD AI-oriented chips. Companies from Hewlett Packard Enterprises HPE to Meta Platforms META and Oracle ORCL are using large quantities of them to build out enterprise-quality HPC applications requiring the lowest latency. Oracle's news includes a supercluster that can link over 16,000 MI300Xs, using a suite of tools provided by AMD to operate them. Hewlett Packard Enterprise will package eight MI325X and two EPYC CPUs into its latest servers, the HPE Proliant XD685, for AI service providers and large model builders. The suite of tools is critical to AMD's success; the ROCm suite is comparable to NVIDIA's CUDA, allowing programmers to harness the power of its GPUs for AI computing needs. The takeaway is that NVIDIA was first to the game. Still, AMD is positioned to catch up and will post accelerating results, outpacing consensus figures and driving shareholder value over the next four to eight quarters. AMD Sets Low Bar With Q3 Guidance Despite 15% Growth Forecast AMD's Q3 guidance is solid but sets a low bar to clear. The company forecasted $6.71 billion in net sales, a 15% gain and an acceleration from the previous quarter but slower than the pace of industry growth and AI peers. AMD growth will be supported by the data center segment, which grew at a triple-digit pace in Q2 and is expected to be sustained in Q3, offset by slower client segment growth and gaming segment normalization. The data center is an increasingly important business segment, growing to nearly 50% of revenue in the year's first half, and contribution gains are expected to continue. Gaming revenue is likely to fall in Q3 compared to last year, but the headwind is diminishing. The pace of contraction should slow in Q3 and Q4 and revert to growth in 2025, becoming a tailwind for the business. Among the risks for AMD investors is government regulation. The U.S. government is considering export bans to risky countries to limit the spread of AI capability among bad actors, potentially limiting the global addressable market. The guidance will drive the market regardless of growth and outperformance in Q3. The consensus for the year implies another 15% gain in revenue and a widening margin, likely underestimating demand for AMD products. Among the calendar Q3 news highlights is the launch of Ryzen AI Pro 300 series chips designed to enhance Microsoft Copilot AI Assist. The chips are 40% faster than the leading competition. They are expected to increase enterprise workload productivity by up to 14%, which is a compelling factor for IT budget managers and business planning. Other highlights include the Turin EPYC data center CPUs that rival NVIDIA's Blackwell, which is expected to launch in Q1 2025. Analysts Eye AMD's Q3 Report as a Catalyst for Recovery Analysts have been raising their estimates for AMD's stock price while lowering estimates for earnings. The consensus is up more than 40% since late 2023 and rising ahead of the Q3 earnings release, forecasting a nearly 20% upside from the $160 level. The latest updates reaffirm targets raised following the Q2 release and put the stock in the $180 to $200 range to align with the consensus. A solid earnings report should catalyze the market to move higher. The price action in AMD stock is tepid following the October investor day and product announcements. The market is falling, down nearly 10% from the recent peak, and may continue to fall until the earnings report is released. The critical support target is near $152.50 and will likely be reached soon, providing an attractive entry point for traders and investors. If market support is not sustained at that level and a lower low is set, a move to $140 is likely, and $120 is possible. The article "AMD Gains Momentum With AI: Can It Beat Expectations?" first appeared on MarketBeat. Market News and Data brought to you by Benzinga APIs
[9]
AMD Believes in Realistic Goals, Not Overpromising
AMD CEO Lisa Su expects the data centre AI accelerator total addressable market (TAM) to grow by more than 60% annually, reaching $500 billion by 2028. AMD is gradually finding its footing in the AI data centre market, emerging as a strong competitor to NVIDIA with its latest AI accelerators and expanding partnerships. At the Advancing 2024 event, the company introduced its new MI325X accelerators for training and inferencing LLMs. The AMD Instinct MI325X accelerators offer leading memory capacity and bandwidth, featuring 256GB of HBM3E with 6.0TB/s throughput -- 1.8x more capacity and 1.3x higher bandwidth than the NVIDIA H200. They also deliver 1.3x higher peak theoretical FP16 and FP8 compute performance than the H200. When asked how AMD compares itself to NVIDIA, Andrew Dieckmann, CVP & GM of data centre GPU at AMD, told AIM that they benchmark against NVIDIA's highest-performing solution. "We are trying to take very representative benchmarks that are realistic. I can tell you that in our customer engagements, especially regarding inference workloads, we have yet to find a single workload that we cannot outperform NVIDIA on," he said, adding that AMD doesn't always outperform NVIDIA. "However, if we optimise for a specific solution, we can beat them." The company also previewed the upcoming AMD Instinct MI350 series accelerators, scheduled for release in 2025, promising a 35x improvement in inference performance over current models and featuring up to 288GB of HBM3E memory. Furthermore, the company plans to launch the MI400 in 2026. Interestingly, on the same day, NVIDIA made headlines by delivering its much-anticipated Blackwell GPUs to OpenAI and Microsoft. Microsoft announced that Azure is the first cloud platform to run NVIDIA's Blackwell system with GB200-powered AI servers. "Our long-standing partnership with NVIDIA and deep innovation continues to lead the industry, powering the most sophisticated AI workloads," said Microsoft CEO Satya Nadella. According to a recent report, NVIDIA's Blackwell GPUs are sold out for the next 12 months, reflecting a similar supply situation that occurred with Hopper GPUs several quarters ago. Consequently, NVIDIA is anticipated to gain market share next year. In the latest quarter, AMD reported revenue of $5.8 billion, while NVIDIA continues to dominate the AI chip market with an impressive quarterly revenue of $30 billion. With NVIDIA's GPUs sold out for the next year, AMD has an ideal opportunity to meet the demand from customers seeking access to compute resources for training and running LLMs. AMD CEO Lisa Su expects the data centre AI accelerator's total addressable market (TAM) to grow by more than 60% annually, reaching $500 billion by 2028. "For AMD, this represents a significant growth opportunity," she said. According to her, AMD GPUs are well-suited for running open-source models like Meta's Llama 3.1 and Stable Diffusion, outperforming NVIDIA's H200. "When you look at that across some of the key models, we're delivering 20 to 40% better inference performance and latency on models like Llama and Mixtral," said Su in her keynote address. "The MI325 platform delivers up to 40% more inference performance than the H200 on Llama 3.1. Many customers are also focused on training, and we've made significant progress in optimising our software stack for training," she added. Moreover, to challenge NVIDIA's CUDA, AMD launched ROCm 6.2, which introduces support for essential AI features such as the FP8 datatype, Flash Attention 3, Kernel Fusion, and more. These updates enable ROCm 6.2 to deliver up to a 2.4X performance boost in inference and a 1.8X improvement in training across a range of LLMs compared to ROCm 6.0. "ROCm is a complete set of libraries, runtime compilers, and tools needed to develop and deploy AI workloads. We designed ROCm to be modular and open-source, allowing for rapid contributions from AI communities," said Vamsi Bopanna, SVP of AI at AMD, adding that it is also designed to connect easily with ecosystem components and frameworks like PyTorch and model hubs like Hugging Face. He explained that they have expanded support for newer frameworks like Jax and implemented powerful new features, algorithms, and optimisations to deliver the best performance for generative AI workloads. AMD also supports various open-source frameworks, including vLLM, Triton, SGlang, and ONXX Runtime and more. Bopanna revealed that today, over 1 million Hugging Face models run on AMD. The company recently acquired European's private AI lab Silo AI. "We recently completed the acquisition of Silo AI, which adds a world class team with tremendous experience training and optimising LLMS and also delivering customer specific AI solutions," said Su. At the event, AMD showcased testimonials for ROCm by inviting startup leaders, including Amit Jain, the CEO of Luma AI; Ashish Vashwani, the CEO of Essential AI; Dani Yogatama, the CEO of Reka AI, and Dmytro Dzhulgakov, the CTO of Fireworks AI. Luma AI recently launched a video generation model called Dream Machine. "The models we're training are very challenging and don't resemble LLMs at all. However, we've been impressed with how quickly we were able to get the model running on ROCm and MI300X GPUs. It took us just a few days to establish the end-to-end pipeline, which is quite fantastic," said Jain. AMD is partnering with customers including Meta, Microsoft, xAI, Oracle, and Cohere, among others. Su highlighted Oracle as a key customer for AMD's latest GPUs. "They've integrated AMD across their entire infrastructure, using our CPUs, GPUs, and DPUs," she said. Oracle SVP Karan Batta joined Su on stage to discuss how Oracle's customers are utilising AMD's hardware tech stack. "Our largest cloud-native customer is Uber. They use Oracle Cloud Infrastructure (OCI) Compute E5 instances with 4th generation AMD EPYC processors to achieve significant performance efficiency. Almost all of their trip-serving infrastructure now runs on AMD within OCI compute," said Batta. "We also have Red Bull Powertrains developing the next generation of F1 engines for upcoming seasons. Additionally, our database franchise is now powered by AMD CPUs. Customers like PayPal and Banco do Brasil are using Exadata powered by AMD to enhance their database portfolios," he added. Alongside Oracle, Databricks is another major customer of AMD. "The large memory capacity and incredible compute capabilities of MI300X have been key to achieving over a 50% increase in performance on some of our critical workloads," said Naveen Rao, VP of generative AI at Databricks, adding that this includes models like Llama and other proprietary models. Microsoft, the first cloud provider to receive NVIDIA Blackwell GPUs, is also partnering with AMD to obtain the new MI325 accelerators. "We're very excited to see how the teams are coming together. OpenAI, Microsoft, and AMD are all working to accelerate the benefits so that this technology can diffuse even faster. We look forward to the roadmap for the MI350 and the next generation after that," said Nadella.
[10]
AMD's Newfound Obsession with 'Feeding the AI Beast'
"To truly feed the AI beast, we must innovate across every component -- GPUs, CPUs, and networking." At Advancing AI 2024, AMD recognised the explosive growth of AI models and the strain it places on both network and computational infrastructure. The company showed its focus on ensuring that every component -- from CPUs, GPUs, and DPUs to networking infrastructure -- is tailored to fuel AI workloads efficiently and at scale. AMD's innovations aim to eliminate any barriers that prevent GPUs from operating at their full potential, enabling faster training and inference cycles, and ultimately creating an ecosystem where AI models can grow without limits. "AMD is the only company that can deliver the full set of CPU, GPU, and networking solutions to address all of the needs of the modern data center. And, we have accelerated our roadmaps to deliver even more innovations across both our Instinct and EPYC portfolios, while also working with an open ecosystem of other leaders to deliver industry-leading networking solutions," said Lisa Su, chair and CEO at AMD. AMD's senior vice president of networking, Soni Jiandani, emphasised the growing demands of AI, stating, "Networking is not only critical, but it's foundational to drive optimal performance." With the launch of the UEC-ready Polara 400 AI networking adapter, AMD promises to elevate performance by ensuring seamless GPU communication across AI clusters, meeting the ever-growing demands of modern AI workloads. This innovation, along with other announcements such as the third-generation Selena DPU, positions AMD to lead the computing and networking industries to the next level, offering up to a six-fold performance improvement in AI training times. "The explosive growth of AI workloads requires innovations across GPUs, CPUs, and networking. AMD is innovating on all fronts to address this growth with our products like Polara 400 and Selena DPU." said Jiandani. Polara 400 AI Networking Adapter: This new networking card is the first UEC-ready AI adapter designed to improve GPU communication, reducing congestion and improving job completion time by intelligently rerouting data across optimal paths Selena DPU: AMD announced the release of its third-generation DPU, the Selena, which offers 400 GB throughput, enabling AI workloads to accelerate with better security, load balancing, and congestion management "To truly feed the AI beast, we must innovate across every component -- GPUs, CPUs, and networking. The Polara 400 and Selena DPU are designed to handle the immense scale and speed required by AI workloads, ensuring the AI beast is always running at its full potential," said Jiandani. Forrest Norrod, EVP and GM at data center solution business at AMD, said that both Selena and Polara will be available early next year. Rajagopal Subramaniyan, SVP of networking at Oracle Cloud Infrastructure (OCI) said that it is leveraging AMD's Elba DPU for high performance and scalability in their AI workloads. "The programmable architecture of Elba of the DPUs has helped us leverage the flexibility to launch features and services for our customers with software-like agility, while maintaining line rate at the hardware. The Elba delivered 5x improvement in performance... which was critical for us to cater to the demands of high-performance requirements of some of our key customers," he added, looking forward to using Selena and Polara. IBM Cloud's Ajay Apte said that it has adopted AMD's DPU for virtualised environments, improving security and performance for enterprise AI and cloud workloads. "By moving our crown jewel, which is the SDN stack, into the DPU, we are essentially increasing the security aspect of our network... with AMD Pensando's programmable DPUs, we are now running our virtualised machine offerings on a more secure and performance-optimized platform," he added. At Advancing AI, Cisco, Microsoft, and AMD announced that they are joining forces to develop a next-generation AI Smart Switch that integrates AMD's programmable DPU into data center switches. This collaboration aims to offload tasks such as security, load balancing, and network traffic management from CPUs and GPUs, significantly improving the performance of AI networks. Microsoft Azure general manager and head of products, Narayan Annamalai highlighted the importance of this development, stating, "We are building an extended network... using DPUs to offload the workload and apply security policies before traffic hits storage or services." "Hypershield is our AI-driven, next-generation security architecture that runs on top of AMD's Pensando DPU inside Cisco servers," added Cisco's senior vice president, Jeremy Foster. This partnership brings AMD's DPU technology into mainstream networking, delivering faster, smarter, and more secure AI infrastructures. By embedding AMD's DPUs into Cisco switches and Microsoft's infrastructure, these companies are ensuring that AI networks can scale efficiently and handle the demands of emerging AI workloads. "Instead of having everything done within the compute servers, we can bring that into the switch over time and implement those technologies... This capability becomes just another capability in your networking fabric across all the switches you are deploying," said SVP and general manager of datacenter and provider connectivity at Cisco, Kevin Wollenweber. AMD is betting big on Ethernet as the foundation for its AI vision, positioning the technology as the most scalable, cost-effective solution for the growing demands of AI workloads. "Ethernet is clearly the preferred choice for back-end and front-end networks for AI workloads... delivering over 50% cost savings and scalability advantages," avered Jiandani. For instance, AMD's new Polara 400 AI networking adapter and its third-generation Selena DPU are at the heart of this Ethernet-driven AI infrastructure, enabling faster data transmission, congestion management, and enhanced scalability across massive GPU clusters. As AI workloads continue to grow exponentially, AMD's Ethernet solutions provide a clear advantage over competitors, allowing for seamless scaling beyond the limitations of traditional architectures like Infiniband. "The foundational architecture of Infiniband is not poised to scale beyond 48,000 nodes without making dramatic and highly complex workarounds, whereas Ethernet has really proven itself to scale to millions of nodes, delivering huge scalability advantages," shared Jiandani, ahead of Advancing AI 2024. AMD believes that its approach to AI networking based on Ethernet helps its customer scale AI clusters, reducing network congestion, and ensuring seamless communication between GPUs. Polara 400 and Selena DPU: These products form the core of AMD's AI networking strategy, using Ethernet to handle increasing data traffic without the complexity and cost associated with alternative networking solutions(AMD Pensando Press Deck). AMD's Ethernet-based approach stands in contrast to competitors like NVIDIA, which relies on solutions like Spectrum-X that incorporate both NICs and switches. By focusing on Ethernet, AMD looks to deliver cost-effective, scalable AI systems without needing additional proprietary hardware. Darrick Horton, CEO of TensorWave, said that it is building large AI clusters with AMD's Polara Ai networking adapters to overcome networking congestion challenges. "With traditional Ethernet, we face challenges with congestion and flow management... Polara Ethernet will allow us to build larger clusters and improve workload efficiency, specifically around job completion times and overall hybrid utilization," he added. Going beyond, AMD is also leading the Ultra Ethernet Consortium (UEC), a coalition of 97 industry-leading vendors, which aims to standardise Ethernet protocols for AI networks, creating a robust ecosystem for high-performance AI and cloud computing. At Advancing AI 2024, AMD also spoke about revolutionising AI systems with its fully programmable P4 engine, designed to accelerate AI networking and offload critical tasks from GPUs and CPUs. Interestingly, this innovative engine is at the core of AMD's Polara 400 and Selena DPU, providing AI systems with the flexibility and performance required to manage the vast amounts of data generated by modern AI workloads. "At the core of this solution powering innovation for both AMD and our customers is our fully programmable P4 engine, which enables us to adapt to the evolution of Ethernet and accommodate AI networking needs," said Jiandani. AMD's P4 engine supports 400-gigabit line-rate throughput and can scale AI networks to handle millions of GPUs, delivering unmatched scalability. "Our fully programmable P4 engine puts AMD in a very unique position because we now have the ability to deliver a holistic portfolio that is future-proof and provides end-to-end network solutions through full programmability," added Jiandani, saying how P4 engine is poised to lead the next generation of AI systems, providing AI workloads with the flexibility to evolve alongside emerging industry standards, outpacing competitors in both performance and adaptability.
[11]
It Took AMD 10 Years to Break into Data Centers, Now It's Poised to Lead AI
What started as a small player with its EPYC processors in 2017 has evolved into a $2.8 billion business, now accounting for 50% of AMD's revenue in Q2 2024. After a decade of relentless innovation, AMD has not only solidified its position in data centers but is also rapidly gaining ground in the AI space. What started as a small player with its EPYC processors in 2017 has evolved into a $2.8 billion business, now accounting for 50% of AMD's revenue in Q2 2024. With its sights set on AI workloads, AMD is set to disrupt the market further, using the same formula that helped it conquer the data center space -- high performance, superior energy efficiency, and competitive TCO. "From virtually zero data center presence in 2017 to a $2.8 billion business in Q2 2024, AMD's rise shows that it's not just a CPU company anymore -- AI is the next frontier," said one of the execs from AMD at the backdrop of Advancing AI 2024. At the keynote speech of AMD Advancing AI 2024, AMD chief Lisa Su said "Since launching in 2017, EPYC has become the CPU of choice for the modern data centre. The largest cloud providers offer more than 950 EPYC instances and have deployed EPYC widely throughout their infrastructure and their most important services -things like Office 365, Facebook, Salesforce, SAP, Zoom, Netflix, and any more." AMD's 5th generation of EPYC processors, particularly the newly launched Turin, are driving a shift in the data center landscape by offering tailored solutions for both cloud computing and AI workloads. The latest processors, built on the 'Zen 5' core architecture, feature up to 192 cores and offer unparalleled speed and efficiency, specifically designed to handle demanding workloads in enterprise, AI, and cloud environments, alongside improving core performance and TCO. "EPYC's energy-efficient processors are not only improving cloud performance but also cutting operational costs by up to 67% compared to older systems, freeing up resources for AI and cloud innovation," said Dan McNamara, senior vice president and general manager, server business, AMD The flagship EPYC 9005 Series processors offer up to 2.7 times the performance of competitors, ensuring AMD maintains its leadership in the data center space. The launch introduces a range of innovations. Notably, the 64-core AMD EPYC 9575F processor, engineered for GPU-powered AI solutions, can boost up to 5GHz, offering a 28% faster data processing speed compared to its closest competitor, whose top speed is 3.8GHz. This performance enhancement is critical for AI workloads, where data must be rapidly fed into GPUs for processing. As McNamara added, "AMD has earned the trust of customers who value demonstrated performance, innovation, and energy efficiency." Source: AMD "AMD has proven it can meet the needs of the data centre market and give customers the standard for performance, efficiency, solutions, and capabilities for cloud, enterprise, and AI workloads," said McNamara. AMD's strategy of on-time roadmap execution has built trust among customers, particularly in sectors where high performance and energy efficiency are critical. At the AMD Advancing AI 2024 event currently underway in San Francisco, USA, AIM noted that Reliance and TATA are leveraging 5th Generation AMD EPYC Processors to fuel India's data centre ambitions. "Our data centre business has gone from virtually zero in 2017, to 50% of the business today, and Q2 was $2.8 billion so all that will begin for the data centre," said McNamara. After a decade of relentless innovation, AMD has not only solidified its position in data centers but is also rapidly gaining ground in the AI space. What started as a small player with its EPYC processors in 2017 has evolved into a $2.8 billion business, now accounting for 50% of AMD's revenue in Q2 2024. With its sights set on AI workloads, AMD is set to disrupt the market further, using the same formula that helped it conquer the data centre space -- high performance, superior energy efficiency, and competitive TCO. The 5th Gen EPYC processors also deliver significant benefits in enterprise workloads, offering up to 17% better instructions per clock (IPC) for cloud and enterprise workloads, and up to 37% higher IPC for AI and high-performance computing (HPC) tasks compared to the previous "Zen 4" architecture. This translates into faster time-to-insights for science and HPC applications and improved performance in virtualized infrastructure, where EPYC processors deliver 1.6 times the performance per core over Intel Xeon servers. In the AI space, the 192-core EPYC 9965 stands out by delivering up to 3.7 times the performance on end-to-end AI workloads, making it a crucial component for generative AI solutions. This processor significantly boosts throughput performance for models like Meta's Llama, making it highly suitable for enterprises deploying generative AI models. With broad support from major hardware providers such as Cisco, Dell, Hewlett Packard Enterprise, Lenovo, and Supermicro, AMD's 5th Gen EPYC processors are set to power the next wave of data center innovation.
[12]
AMD unveils AI-infused chips across Ryzen, Instinct and Epyc brands
Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Speaking at an event in San Francisco, AMD CEO Lisa Su unveiled AI-infused chips across its Ryzen, Instinct and Epyc brands, fueling a new generation of AI computing for everyone from business users to data centers. Throughout the event, AMD indirectly made references to rivals such as Nvidia and Intel by emphasizing its quest to provide technology that was open and accessible to the widest variety of customers, without an intent to lock those customers into proprietary solutions. Su said AI will boost our personal productivity, collaboration will become much better with things like real-time translate, and it will make life easier whether you are a creator or ordinary user. It will be processed locally, to protect your privacy, Su said. She noted the new AMD Ryzen AI Pro PCs will be CoPilot+-ready and offer up to 23 hours of battery life (and nine hours using Microsoft Teams). "We've been working very closely with AI PC ecosystem developers," she said, noting more than 100 will be working on AI apps by the end of the year. Commercial AI mobile Ryzen processors AMD announced its third generation commercial AI mobile processors, designed specifically to transform business productivity with Copilot+ features including live captioning and language translation in conference calls and advanced AI image generators. If you really wanted to, you could use AI-based Microsoft Teams for up to nine hours on new laptops equipped with the AMD processors. The new Ryzen AI PRO 300 Series processors deliver industry-leading AI compute, with up to three times the AI performance than the previous generation of AMD processors. More than 100 products using the Ryzen processors are on the way through 2025. Enabled with AMD PRO Technologies, the Ryzen AI PRO 300 Series processors offer high security and manageability features designed to streamline IT operations and ensure exceptional ROI for businesses. Ryzen AI PRO 300 Series processors feature new AMD Zen 5 architecture, delivering outstanding CPU performance, and are the world's best line up of commercial processors for Copilot+ enterprise PCs5. Zen, now in its fifth generation, has been the foundation behind AMD's own financial recovery, its gains in market share against Intel, and Intel's own subsequent hard times and layoffs. "I think the best is that AMD continue to execute on a solid product roadmap. Unfortunately they are making performance comparisons to the competition's previous generation products," said Jim McGregor, an analyst at Tirias Research, in an email to VentureBeat. "So, we have to wait and see how the products will compare. However, I do expect them to be highly competitive especially the processors. Note that AMD only announced a new architecture for nenetworking, everything else is evolutionary but that's not a bad thing when you are in a strong position and gaining market share." Laptops equipped with Ryzen AI PRO 300 Series processors are designed to tackle business' toughest workloads, with the top-of-stack Ryzen AI 9 HX PRO 375 offering up to 40% higher performance and up to 14% faster productivity performance compared to Intel's Core Ultra 7 165U, AMD said. With the addition of XDNA 2 architecture powering the integrated NPU (the neural processing unit, or AI-focused part of the processor), AMD Ryzen AI PRO 300 Series processors offer a cutting-edge 50+ NPU TOPS (Trillions of Operations Per Second) of AI processing power, exceeding Microsoft's Copilot+ AI PC requirements and delivering exceptional AI compute and productivity capabilities for the modern business. Built on a 4 nanometer (nm) process and with innovative power management, the new processors deliver extended battery life ideal for sustained performance and productivity on the go. "Enterprises are increasingly demanding more compute power and efficiency to drive their everyday tasks and most taxing workloads. We are excited to add the Ryzen AI PRO 300 Series, the most powerful AI processor built for business PCs10 , to our portfolio of mobile processors," said Jack Huynh, senior vice president and general manager of the computing and graphics group at AMD, in a statement. "Our third generation AI-enabled processors for business PCs deliver unprecedented AI processing capabilities with incredible battery life and seamless compatibility for the applications users depend on." AMD expands commercial OEM ecosystem OEM partners continue to expand their commercial offerings with new PCs powered by Ryzen AI PRO 300 Series processors, delivering well-rounded performance and compatibility to their business customers. With industry leading TOPS, the next generation of Ryzen processor-powered commercial PCs are set to expand the possibilities of local AI processing with Microsoft Copilot+. OEM systems powered by Ryzen AI PRO 300 Series are expected to be on shelf starting later this year. "Microsoft's partnership with AMD and the integration of Ryzen AI PRO processors into Copilot+ PCs demonstrate our joint focus on delivering impactful AI-driven experiences for our customers. The Ryzen AI PRO's performance, combined with the latest features in Windows 11, enhances productivity, efficiency, and security," said Pavan Davuluri, corporate vice president for Windows+ Devices at Microsoft, in a statement. "Features like Improved Windows Search, Recall, and Click to Do make PCs more intuitive and responsive. Security enhancements, including the Microsoft Pluton security processor and Windows Hello Enhanced Sign-in Security, help safeguard customer data with advanced protection. We're proud of our strong history of collaboration with AMD and are thrilled to bring these innovations to market." "In today's AI-powered era of computing, HP is dedicated to delivering powerful innovation and performance that revolutionizes the way people work," said Alex Cho, president of Personal Systems at HP, in a statement. "With the HP EliteBook X Next-Gen AI PC, we are empowering modern leaders to push boundaries without compromising power or performance. We are proud to expand our AI PC lineup powered by AMD, providing our commercial customers with a truly personalized experience." "Lenovo's partnership with AMD continues to drive AI PC innovation and deliver supreme performance for our business customers. Our recently announced ThinkPad T14s Gen 6 AMD, powered by the latest AMD Ryzen AI PRO 300 Series processors, showcases the strength of our collaboration," said Luca Rossi, president, Lenovo Intelligent Devices Group. "This device offers outstanding AI computing power, enhanced security, and exceptional battery life, providing professionals with the tools they need to maximize productivity and efficiency. Together with AMD, we are transforming the business landscape by delivering smarter, AIdriven solutions that empower users to achieve more." New Pro Technologies features for security and management In addition to AMD Secure Processor, AMD Shadow Stack and AMD Platform Secure Boot, AMD has expanded its Pro Technologies lineup with new security and manageability features. Processors equipped with PRO Technologies will now come standard with Cloud Bare Metal Recovery, allowing IT teams to seamlessly recover systems via the cloud ensuring smooth and continuous operations; Supply Chain Security (AMD Device Identity), a new supply chain security function, enabling traceability across the supply chain; and Watch Dog Timer, building on existing resiliency support with additional detection and recovery processes. Additional AI-based malware detection is available via PRO Technologies with select ISV partners. These new security features leverage the integrated NPU to run AI-based security workloads without impacting day-to-day performance. AMD unveils Instinct MI325X accelerators for AI data centers AMD has become a big player in the graphics processing units (GPUs) for data centers, and today it announced the latest AI accelerators and networking solutions for AI infrastructure. The company unveiled the AMD Instinct MI325X accelerators, the AMD Pensando Pollara 400 network interface card (NIC) and the AMD Pensando Salina data processing unit (DPU). AMD claimed the AMD Instinct MI325X accelerators set a new standard in performance for Gen AI models and data centers. Built on the AMD CDNA 3 architecture, AMD Instinct MI325X accelerators are designed for performance and efficiency for demanding AI tasks spanning foundation model training, fine-tuning and inferencing. Together, these products enable AMD customers and partners to create highly performant and optimized AI solutions at the system, rack and data center level. "AMD continues to deliver on our roadmap, offering customers the performance they need and the choice they want, to bring AI infrastructure, at scale, to market faster," said Forrest Norrod, executive vice president and general manager of the data center solutions business group at AMD, in a statement. "With the new AMD Instinct accelerators, EPYC processors and AMD Pensando networking engines, the continued growth of our open software ecosystem, and the ability to tie this all together into optimized AI infrastructure, AMD underscores the critical expertise to build and deploy world class AI solutions." AMD Instinct MI325X accelerators deliver industry-leading memory capacity and bandwidth, with 256GB of HBM3E supporting 6.0TB/s offering 1.8 times more capacity and 1.3 times more bandwidth than the Nvidia H200, AMD said. The AMD Instinct MI325X also offers 1.3 times greater peak theoretical FP16 and FP8 compute performance compared to H200. This leadership memory and compute can provide up to 1.3 times the inference performance on Mistral 7B at FP162, 1.2 times the inference performance on Llama 3.1 70B at FP83 and 1.4 times the inference performance on Mixtral 8x7B at FP16 of the H200. (Nvidia has more recent devices on the market now and they are not yet available for comparisons, AMD said). "AMD certainly remains well positioned in the data center, but I think their CPU efforts are still their best positioned products. The market for AI accelleration/GPUs is still heavily favoring Nvidia and I don't see that changing anytime soon. But the need for well optimized and purpose designed CPUs to compliment as a host processor any AI accelerator or GPU is essential and AMDs datacenter CPUs are competitive there," said Ben Bajarin, an analyst at Creative Strategies, in an email to VentureBeat. "On the networking front, there is certainly good progress here technically and I imagine the more AMD can integrate this into their full stack approach to optimizing for the racks via the ZT systems purchase, then I think their networking stuff becomes even more important." He added, "Broad point to make here, is the data center is under a complete transformation and we are still only in the early days of that which makes this still a wide open competitive field over the arc of time 10+ years. I'm not sure we can say with any certainty how this shakes out over that time but the bottom line is there is a lot of market share and $$ to go around to keep AMD, Nvidia, and Intel busy." AMD Instinct MI325X accelerators are currently on track for production shipments in Q4 2024 and are expected to have widespread system availability from a broad set of platform providers, including Dell Technologies, Eviden, Gigabyte, Hewlett Packard Enterprise, Lenovo, Supermicro and others starting in Q1 2025. Updating its annual roadmap, AMD previewed the next-generation AMD Instinct MI350 series accelerators. Based on the AMD CDNA 4 architecture, AMD Instinct MI350 series accelerators are designed to deliver a 35 times improvement in inference performance compared to AMD CDNA 3-based accelerators. The AMD Instinct MI350 series will continue to drive memory capacity leadership with up to 288GB of HBM3E memory per accelerator. The AMD Instinct MI350 series accelerators are on track to be available during the second half of 2025. "AMD undoubtedly increased the distance between itself and Intel with Epyc. It currently has 50-60% market share with the hyoerscalers and I don't see that abating. AMD;'s biggest challenge is to get share with enterprises. Best product rarely wins in the enterprise and AMD needs to invest more into sales and marketing to accelerate its enterprise growth," said Patrick Moorhead, an analyst at Moor Insights & Strategy, in an email to VentureBeat. "It's s bit harder to assess where AMD sits versus NVIDIA in Datacenter GPUs. There's numbers flying all around, claims from both companies that they're better. Signal65, our sister benchmarking company, hasn't had the opportunity to do our own tests." And Moohead added, "What I can unequivocally say is that AMD's new GPUs, particularly the MI350, is a massive improvement given improved efficiency, performance and better support for lower bit rate models than its predecessors. It is a two horse race, with Nvidia in the big lead and AMD is quickly catching up and providing meaningful results. The facts that Meta's live llama 405B model runs exclusively on MI is a huge statement on competitiveness. " AMD next-gen AI Networking AMD is leveraging the most widely deployed programmable DPU for hyperscalers to power next-gen AI networking, said Soni Jiandani, senior vice president of the network technology solutions group, in a press briefing. Split into two parts: the front-end, which delivers data and information to an AI cluster, and the backend, which manages data transfer between accelerators and clusters, AI networking is critical to ensuring CPUs and accelerators are utilized efficiently in AI infrastructure. To effectively manage these two networks and drive high performance, scalability and efficiency across the entire system, AMD introduced the AMD Pensando Salina DPU for the front-end and the AMD Pensando Pollara 400, the industry's first Ultra Ethernet Consortium (UEC) ready AI NIC, for the back-end. The AMD Pensando Salina DPU is the third generation of the world's most performant and programmable DPU, bringing up to two times the performance, bandwidth and scale compared to the previous generation. Supporting 400G throughput for fast data transfer rates, the AMD Pensando Salina DPU is a critical component in AI front-end network clusters, optimizing performance, efficiency, security and scalability for data-driven AI applications. The UEC-ready AMD Pensando Pollara 400, powered by the AMD P4 Programmable engine, is the industry's first UEC-ready AI NIC. It supports the next-gen RDMA software and is backed by an open ecosystem of networking. The AMD Pensando Pollara 400 is critical for providing leadership performance, scalability and efficiency of accelerator-to-accelerator communication in back-end networks. Both the AMD Pensando Salina DPU and AMD Pensando Pollara 400 are sampling with customers in Q4'24 and are on track for availability in the first half of 2025. AMD AI software for Generative AI AMD continues its investment in driving software capabilities and the open ecosystem to deliver powerful new features and capabilities in the AMD ROCm open software stack. Within the open software community, AMD is driving support for AMD compute engines in the most widely used AI frameworks, libraries and models including PyTorch, Triton, Hugging Face and many others. This work translates to out-of-the-box performance and support with AMD Instinct accelerators on popular generative AI models like Stable Diffusion 3, Meta Llama 3, 3.1 and 3.2 and more than one million models at Hugging Face. Beyond the community, AMD continues to advance its ROCm open software stack, bringing the latest features to support leading training and inference on Generative AI workloads. ROCm 6.2 now includes support for critical AI features like FP8 datatype, Flash Attention 3, Kernel Fusion and more. With these new additions, ROCm 6.2, compared to ROCm 6.0, provides up to a 2.4X performance improvement on inference6 and 1.8X on training for a variety of LLMs. AMD launches 5th Gen AMD Epyc CPUs for the data center AMD also announced the availability of the 5th Gen AMD Epyc processors, formerly codenamed "Turin," the "world's best server CPU for enterprise, AI and cloud," the company said. Using the Zen 5 core architecture, compatible with the broadly deployed SP5 platform and offering a broad range of core counts spanning from eight to 192, the AMD Epyc 9005 Series processors extend the record-breaking performance and energy efficiency of the previous generations with the top of stack 192 core CPU delivering up to 2.7 times the performance compared to the competition, AMD said. New to the AMD Epyc 9005 Series CPUs is the 64 core AMD Epyc 9575F, tailor made for GPU-powered AI solutions that need the ultimate in host CPU capabilities. Boosting up to 5GHz, compared to the 3.8GHz processor of the competition, it provides up to 28% faster processing needed to keep GPUs fed with data for demanding AI workloads, AMD said. "From powering the world's fastest supercomputers, to leading enterprises, to the largest Hyperscalers, AMD has earned the trust of customers who value demonstrated performance, innovation and energy efficiency," said Dan McNamara, senior vice president and general manager of the server business at AMD, in a statement. "With five generations of on-time roadmap execution, AMD has proven it can meet the needs of the data center market and give customers the standard for data center performance, efficiency, solutions and capabilities for cloud, enterprise and AI workloads." In a press briefing, McNamara thanked Zen for AMD's server market share rise from zero in 2017 to 34% in the second quarter of 2024 (according to Mercury Research). Modern data centers run a variety of workloads, from supporting corporate AI-enablement initiatives, to powering large-scale cloud-based infrastructures to hosting the most demanding business-critical applications. The new 5th Gen AMD Epyc processors provide leading performance and capabilities for the broad spectrum of server workloads driving business IT today. "This is a beast," McNamara said. "We are really excited about it." The new Zen 5 core architecture, provides up to 17% better instructions per clock (IPC) for enterprise and cloud workloads and up to 37% higher IPC in AI and high performance computing (HPC) compared to Zen 4. With AMD Epyc 9965 processor-based servers, customers can expect significant impact in their real world applications and workloads compared to the Intel Xeon 8592+ CPU-based servers, with: up to four times faster time to results on business applications such as video transcoding. AMD said it also has up to 3.9 times the time to insights for science and HPC applications that solve the world's most challenging problems; up to 1.6 times the performance per core in virtualized infrastructure. In addition to leadership performance and efficiency in general purpose workloads, the 5th Gen AMD Epyc processors enable customers to drive fast time to insights and deployments for AI deployments, whether they are running a CPU or a CPU + GPU solution, McNamara said. Compared to the competition, he said the 192 core Epyc 9965 CPU has up to 3.7 times the performance on end-to-end AI workloads, like TPCx-AI (derivative), which are critical for driving an efficient approach to generative AI. In small and medium size enterprise-class generative AI models, like Meta's Llama 3.1-8B, the Epyc 9965 provides 1.9 times the throughput performance compared to the competition. Finally, the purpose built AI host node CPU, the EPYC 9575F, can use its 5GHz max frequency boost to help a 1,000 node AI cluster drive up to 700,000 more inference tokens per second. Accomplishing more, faster. By modernizing to a data center powered by these new processors to achieve 391,000 units of SPECrate2017_int_base general purpose computing performance, customers receive impressive performance for various workloads, while gaining the ability to use an estimated 71% less power and ~87% fewer servers. This gives CIOs the flexibility to either benefit from the space and power savings or add performance for day-to-day IT tasks while delivering impressive AI performance. The entire lineup of 5th Gen AMD EPYC processors is available today, with support from Cisco, Dell, Hewlett Packard Enterprise, Lenovo and Supermicro as well as all major ODMs and cloud service providers providing a simple upgrade path for organizations seeking compute and AI leadership. Dell said that said its 16-accelerated PowerEdge servers would be able to replace seven prior generation servers, with a 65% reduction of energy usage. HP Enterprise also took the stage to say Lumi, one of its customers, is working on a digital twin of the entire planet, dubbed Destination Earth, using the AMD tech. Daniel Newman, CEO of The Futurum Group and an analyst, said in an email to VentureBeat, "Instinct and the new MI325X will be the hot button from today's event. It isn't a completely new launch, but the Q4 ramp will run alongside nvidia Blackwell and will be the next important indicator of AMD's trajectory as the most compelling competitor to Nvidia. The 325X ramping while the new 350 will be the biggest leap when it launches in 2H of 2025 making a 35 times AI performance leap from its CDNA3. " Newman added, "Lisa Su's declaration of a $500 billion AI accelerator market between 2023 and 2028 is an incredibly ambitious leap that represents more than 2x our current forecast and indicates a material upside for the market coming from a typically conservative CEO in Lisa Su. Other announcements in networking and compute (Turin) show the company's continued expansion and growth." And he said, "The Epyc DC CPU business showed significant generational improvements. AMD has been incredibly successful in winning cloud datacenter business for its EPYC line now having more than 50% of share and in some cases we believe closer to 80%. For AMD, the big question is can it turn the strength in cloud and turn its attention to enterprise data center where Intel is still dominant-this could see AMD DC CPU business expand to more than its already largest ever 34%. Furthermore, can the company take advantage of its strength in cloud to win more DC GPU deals and fend off NVIDIA's strength at more than 90% market share."
[13]
AMD Unveils Leadership AI Solutions at Advancing AI 2024 - Advanced Micro Devices (NASDAQ:AMD)
-- AMD launches 5th Gen AMD EPYC processors, AMD Instinct MI325X accelerators, next gen networking solutions and AMD Ryzen AI PRO processors powering enterprise AI at scale -- -- Dell, Google Cloud, HPE, Lenovo, Meta, Microsoft, Oracle Cloud Infrastructure, Supermicro and AI leaders Databricks, Essential AI, Fireworks AI, Luma AI and Reka AI joined AMD to showcase expanding AMD AI solutions for enterprises and end users -- -- Technical leaders from Cohere, Google DeepMind, Meta, Microsoft, OpenAI and more discussed how they are using AMD ROCm software to deploy models and applications on AMD Instinct accelerators -- SAN FRANCISCO, Oct. 10, 2024 (GLOBE NEWSWIRE) -- AMD AMD today launched the latest high performance computing solutions defining the AI computing era, including 5th Gen AMD EPYC™ server CPUs, AMD Instinct™ MI325X accelerators, AMD Pensando™ Salina DPUs, AMD Pensando Pollara 400 NICs and AMD Ryzen™ AI PRO 300 series processors for enterprise AI PCs. AMD and its partners also showcased how they are deploying AMD AI solutions at scale, the continued ecosystem growth of AMD ROCm™ open source AI software, and a broad portfolio of new solutions based on AMD Instinct accelerators, EPYC CPUs and Ryzen PRO CPUs. "The data center and AI represent significant growth opportunities for AMD, and we are building strong momentum for our EPYC and AMD Instinct processors across a growing set of customers," said AMD Chair and CEO Dr. Lisa Su. "With our new EPYC CPUs, AMD Instinct GPUs and Pensando DPUs we are delivering leadership compute to power our customers' most important and demanding workloads. Looking ahead, we see the data center AI accelerator market growing to $500 billion by 2028. We are committed to delivering open innovation at scale through our expanded silicon, software, network and cluster-level solutions." Defining the Data Center in the AI Era AMD announced a broad portfolio of data center solutions for AI, enterprise, cloud and mixed workloads: New AMD EPYC 9005 Series processors deliver record-breaking performance1 to enable optimized compute solutions for diverse data center needs. Built on the latest "Zen 5" architecture, the lineup offers up to 192 cores and will be available in a wide range of platforms from leading OEMs and ODMs starting today.AMD continues executing its annual cadence of AI accelerators with the launch of AMD Instinct MI325X, delivering leadership performance and memory capabilities for the most demanding AI workloads. AMD also shared new details on next-gen AMD Instinct MI350 series accelerators expected to launch in the second half of 2025, extending AMD Instinct leadership memory capacity and generative AI performance. AMD has made significant progress developing the AMD Instinct MI400 Series accelerators based on the AMD CDNA Next architecture, planned to be available in 2026.AMD has continuously improved its AMD ROCm software stack, doubling AMD Instinct MI300X accelerator inferencing and training performance2 across a wide range of the most popular AI models. Today, over one million models run seamlessly out of the box on AMD Instinct, triple the number available when MI300X launched, with day-zero support for the most widely used models.AMD also expanded its high performance networking portfolio to address evolving system networking requirements for AI infrastructure, maximizing CPU and GPU performance to deliver performance, scalability and efficiency across the entire system. The AMD Pensando Salina DPU delivers a high performance front-end network for AI systems, while the AMD Pensando Pollara 400, the first Ultra Ethernet Consortium ready NIC, reduces the complexity of performance tuning and helps improve time to production. AMD partners detailed how they leverage AMD data center solutions to drive leadership generative AI capabilities, deliver cloud infrastructure used by millions of people daily and power on-prem and hybrid data centers for leading enterprises: Since launching in December 2023, AMD Instinct MI300X accelerators have been deployed at scale by leading cloud, OEM and ODM partners and are serving millions of users daily on popular AI models, including OpenAI's ChatGPT, Meta Llama and over one million open source models on the Hugging Face platform.Google highlighted how AMD EPYC processors power a wide range of instances for AI, high performance, general purpose and confidential computing, including their AI Hypercomputer, a supercomputing architecture designed to maximize AI ROI. Google also announced EPYC 9005 Series-based VMs will be available in early 2025.Oracle Cloud Infrastructure shared how it leverages AMD EPYC CPUs, AMD Instinct accelerators and Pensando DPUs to deliver fast, energy efficient compute and networking infrastructure for customers like Uber, Red Bull Powertrains, PayPal and Fireworks AI. OCI announced the new E6 compute platform powered by EPYC 9005 processors.Databricks highlighted how its models and workflows run seamlessly on AMD Instinct and ROCm and disclosed that their testing shows the large memory capacity and compute capabilities of AMD Instinct MI300X GPUs help deliver an over 50% increase in performance on Llama and Databricks proprietary models.Microsoft CEO Satya Nadella highlighted Microsoft's longstanding collaboration and co-innovation with AMD across its product offerings and infrastructure, with MI300X delivering strong performance on Microsoft Azure and GPT workloads. Nadella and Su also discussed the companies' deep partnership on the AMD Instinct roadmap and how Microsoft is planning to leverage future generations of AMD Instinct accelerators including MI350 series and beyond to deliver leadership performance-per-dollar-per-watt for AI applications.Meta detailed how AMD EPYC CPUs and AMD Instinct accelerators power its compute infrastructure across AI deployments and services, with MI300X serving all live traffic on Llama 405B. Meta is also partnering with AMD to optimize AI performance from silicon, systems, and networking to software and applications.Leading OEMs Dell, HPE, Lenovo and Supermicro are expanding on their highly performant, energy efficient AMD EPYC processor-based lineups with new platforms designed to modernize data centers for the AI era. Expanding an Open AI Ecosystem AMD continues to invest in the open AI ecosystem and expand the AMD ROCm open source software stack with new features, tools, optimizations and support to help developers extract the ultimate performance from AMD Instinct accelerators and deliver out-of-the-box support for today's leading AI models. Leaders from Essential AI, Fireworks AI, Luma AI and Reka AI discussed how they are optimizing models across AMD hardware and software. AMD also hosted a developer event joined by technical leaders from across the AI developer ecosystem, including Microsoft, OpenAI, Meta, Cohere, xAI and more. Luminary presentations hosted by the inventors of popular AI programming languages, models and frameworks critical to the AI transformation taking place, such as Triton, TensorFlow, vLLM and Paged Attention, FastChat and more, shared how developers are unlocking AI performance optimizations through vendor agnostic programming languages, accelerating models on AMD Instinct accelerators, and highlighted the ease of use porting to ROCm software and how the ecosystem is benefiting from an open-source approach. Enabling Enterprise Productivity with AI PCs AMD launched AMD Ryzen AI PRO 300 Series processors, powering the first Microsoft Copilot+ laptops enabled for the enterprise3. The Ryzen AI PRO 300 Series processor lineup extends AMD leadership in performance and battery life with the addition of enterprise-grade security and manageability features for business users. The Ryzen AI PRO 300 Series processors, featuring the new AMD "Zen 5" and AMD XDNA™ 2 architectures, are the world's most advanced commercial processors4, offering best in class performance for unmatched productivity5 and an industry leading 55 NPU TOPS6 of AI performance with the Ryzen AI 9 HX PRO 375 processor to process AI tasks locally on Ryzen AI PRO laptops.Microsoft highlighted how Windows 11 Copilot+ and the Ryzen AI PRO 300 lineup are ready for next generation AI experiences, including new productivity and security features.OEM partners including HP and Lenovo are expanding their commercial offerings with new PCs powered by Ryzen AI PRO 300 Series processors, with more than 100 platforms expected to come to market through 2025. Supporting Resources Watch the AMD Advancing AI keynote and see the news hereFollow AMD on XConnect with AMD on LinkedIn About AMD For more than 50 years AMD has driven innovation in high-performance computing, graphics, and visualization technologies. Billions of people, leading Fortune 500 businesses, and cutting-edge scientific research institutions around the world rely on AMD technology daily to improve how they live, work, and play. AMD employees are focused on building leadership high-performance and adaptive products that push the boundaries of what is possible. For more information about how AMD is enabling today and inspiring tomorrow, visit the AMD AMD website, blog, LinkedIn, and X pages. Cautionary Statement This press release contains forward-looking statements concerning Advanced Micro Devices, Inc. (AMD) such as the features, functionality, performance, availability, timing and expected benefits of AMD products; AMD's expected data center and AI growth opportunities; the ability of AMD to build momentum for AMD EPYC™ and AMD Instinct™ processors across its customers; the ability of AMD to deliver leadership compute to power to its customers workloads; the anticipated growth of the data center AI accelerator market by 2028; and AMD's commitment to delivering open innovation at scale, which are made pursuant to the Safe Harbor provisions of the Private Securities Litigation Reform Act of 1995. Forward-looking statements are commonly identified by words such as "would," "may," "expects," "believes," "plans," "intends," "projects" and other terms with similar meaning. Investors are cautioned that the forward-looking statements in this press release are based on current beliefs, assumptions and expectations, speak only as of the date of this press release and involve risks and uncertainties that could cause actual results to differ materially from current expectations. Such statements are subject to certain known and unknown risks and uncertainties, many of which are difficult to predict and generally beyond AMD's control, that could cause actual results and other future events to differ materially from those expressed in, or implied or projected by, the forward-looking information and statements. Material factors that could cause actual results to differ materially from current expectations include, without limitation, the following: Intel Corporation's dominance of the microprocessor market and its aggressive business practices; Nvidia's dominance in the graphics processing unit market and its aggressive business practices; the cyclical nature of the semiconductor industry; market conditions of the industries in which AMD products are sold; loss of a significant customer; competitive markets in which AMD's products are sold; economic and market uncertainty; quarterly and seasonal sales patterns; AMD's ability to adequately protect its technology or other intellectual property; unfavorable currency exchange rate fluctuations; ability of third party manufacturers to manufacture AMD's products on a timely basis in sufficient quantities and using competitive technologies; availability of essential equipment, materials, substrates or manufacturing processes; ability to achieve expected manufacturing yields for AMD's products; AMD's ability to introduce products on a timely basis with expected features and performance levels; AMD's ability to generate revenue from its semi-custom SoC products; potential security vulnerabilities; potential security incidents including IT outages, data loss, data breaches and cyberattacks; uncertainties involving the ordering and shipment of AMD's products; AMD's reliance on third-party intellectual property to design and introduce new products; AMD's reliance on third-party companies for design, manufacture and supply of motherboards, software, memory and other computer platform components; AMD's reliance on Microsoft and other software vendors' support to design and develop software to run on AMD's products; AMD's reliance on third-party distributors and add-in-board partners; impact of modification or interruption of AMD's internal business processes and information systems; compatibility of AMD's products with some or all industry-standard software and hardware; costs related to defective products; efficiency of AMD's supply chain; AMD's ability to rely on third party supply-chain logistics functions; AMD's ability to effectively control sales of its products on the gray market; long-term impact of climate change on AMD's business; impact of government actions and regulations such as export regulations, tariffs and trade protection measures; AMD's ability to realize its deferred tax assets; potential tax liabilities; current and future claims and litigation; impact of environmental laws, conflict minerals related provisions and other laws or regulations; evolving expectations from governments, investors, customers and other stakeholders regarding corporate responsibility matters; issues related to the responsible use of AI; restrictions imposed by agreements governing AMD's notes, the guarantees of Xilinx's notes and the revolving credit agreement; impact of acquisitions, joint ventures and/or investments on AMD's business and AMD's ability to integrate acquired businesses; impact of any impairment of the combined company's assets; political, legal and economic risks and natural disasters; future impairments of technology license purchases; AMD's ability to attract and retain qualified personnel; and AMD's stock price volatility. Investors are urged to review in detail the risks and uncertainties in AMD's Securities and Exchange Commission filings, including but not limited to AMD's most recent reports on Forms 10-K and 10-Q. AMD, the AMD Arrow logo, EPYC, AMD CDNA, AMD Instinct, Pensando, ROCm, Ryzen, and combinations thereof are trademarks of Advanced Micro Devices, Inc. Other names are for informational purposes only and may be trademarks of their respective owners. __________________________________ 1 EPYC-022F: For a complete list of world records see: http://amd.com/worldrecords. 2 Testing conducted by internal AMD Performance Labs as of September 29, 2024 inference performance comparison between ROCm 6.2 software and ROCm 6.0 software on the systems with 8 AMD Instinct™ MI300X GPUs coupled with Llama 3.1-8B, Llama 3.1-70B, Mixtral-8x7B, Mixtral-8x22B, and Qwen 72B models. ROCm 6.2 with vLLM 0.5.5 performance was measured against the performance with ROCm 6.0 with vLLM 0.3.3, and tests were performed across batch sizes of 1 to 256 and sequence lengths of 128 to 2048. Configurations: 1P AMD EPYC™ 9534 CPU server with 8x AMD Instinct™ MI300X (192GB, 750W) GPUs, Supermicro AS-8125GS-TNMR2, NPS1 (1 NUMA per socket), 1.5 TiB (24 DIMMs, 4800 mts memory, 64 GiB/DIMM), 4x 3.49TB Micron 7450 storage, BIOS version: 1.8, , ROCm 6.2.0-00, vLLM 0.5.5, PyTorch 2.4.0, Ubuntu® 22.04 LTS with Linux kernel 5.15.0-119-generic. vs. 1P AMD EPYC 9534 CPU server with 8x AMD Instinct™ MI300X (192GB, 750W) GPUs, Supermicro AS-8125GS-TNMR2, NPS1 (1 NUMA per socket), 1.5TiB 24 DIMMS, 4800 mts memory, 64 GiB/DIMM), 4x 3.49TB Micron 7450 storage, BIOS version: 1.8, ROCm 6.0.0-00, vLLM 0.3.3, PyTorch 2.1.1, Ubuntu 22.04 LTS with Linux kernel 5.15.0-119-generic. MI300-62 Server manufacturers may vary configurations, yielding different results. Performance may vary based on factors including but not limited to different versions of configurations, vLLM, and drivers. 3 Based on Microsoft Copilot+ requirements of minimum 40 TOPS using AMD product specifications and competitive products announced as of Oct 2024. Microsoft requirements found here - https://support.microsoft.com/en-us/topic/copilot-pc-hardware-requirements-35782169-6eab-4d63-a5c5-c498c3037364. STXP-05. 4 Based on a small node size for an x86 platform and cutting-edge, interconnected technologies, as of September 2024. GD-203b 5 Testing as of Sept 2024 by AMD performance labs using the following systems: HP EliteBook X G1a with AMD Ryzen AI 9 HX PRO 375 processor @40W, Radeon™ 890M graphics, 32GB of RAM, 512GB SSD, VBS=ON, Windows 11 Pro; Lenovo ThinkPad T14s Gen 6 with AMD Ryzen™ AI 7 PRO 360 processor @22W, Radeon™ 880M graphics, 32GB RAM, 1TB SSD, VBS=ON, Windows 11 Pro; Dell Latitude 7450 with Intel Core Ultra 7 165U processor @15W (vPro enabled), Intel Iris Xe Graphics, VBS=ON, 32GB RAM, 512GB NVMe SSD, Microsoft Windows 11 Professional; Dell Latitude 7450 with Intel Core Ultra 7 165H processor @28W (vPro enabled), Intel Iris Xe Graphics, VBS=ON, 16GB RAM, 512GB NVMe SSD, Microsoft Windows 11 Pro. The following applications were tested in Balanced Mode: Teams + Procyon Office Productivity, Teams + Procyon Office Productivity Excel, Teams + Procyon Office Productivity Outlook, Teams + Procyon Office Productivity Power Point, Teams + Procyon Office Productivity Word, Composite Geomean Score. Each Microsoft Teams call consists of 9 participants (3X3). Laptop manufactures may vary configurations yielding different results. STXP-10. Testing as of Sept 2024 by AMD performance labs using the following systems: (1) Lenovo ThinkPad T14s Gen 6 with an AMD Ryzen™ AI 7 PRO 360 processor (@22W), Radeon™ 880M graphics, 32GB RAM, 1TB SSD, VBS=ON, Windows 11 Pro; (2) Dell Latitude 7450 with Intel Core Ultra 7 165U processor (@15W) (vPro enabled), Intel Iris Xe Graphics, VBS=ON, 32GB RAM, 512GB NVMe SSD, Microsoft Windows 11 Professional; and (3) Dell Latitude 7450 with Intel Core Ultra 7 165H processor (@28W) (vPro enabled), Intel Arc Graphics, VBS=ON, 16GB RAM, 512GB NVMe SSD, Microsoft Windows 11 Pro. Tested applications (in Balanced Mode) include: Procyon Office Productivity, Procyon Office Productivity Excel, Procyon Office Productivity Outlook, Procyon Office Productivity Power Point, Procyon Office Productivity Word, Composite Geomean Score. Laptop manufactures may vary configurations yielding different results. STXP-11. 6 Trillions of Operations per Second (TOPS) for an AMD Ryzen processor is the maximum number of operations per second that can be executed in an optimal scenario and may not be typical. TOPS may vary based on several factors, including the specific system configuration, AI model, and software version. GD-243. Media Contacts: Brandi Martina AMD Communications +1 512-705-1720 brandi.martina@amd.com Mitch Haws AMD Investor Relations +1 512-944-0790 mitch.haws@amd.com Market News and Data brought to you by Benzinga APIs
[14]
AMD Delivers Leadership AI Performance with AMD Instinct MI325X Accelerators - Advanced Micro Devices (NASDAQ:AMD)
-- Latest accelerators offer market leading HBM3E memory capacity and are supported by partners and customers including Dell Technologies, HPE, Lenovo, Supermicro and others -- -- AMD Pensando Salina DPU offers 2X generational performance and AMD Pensando Pollara 400 is industry's first UEC ready NIC-- SAN FRANCISCO, Oct. 10, 2024 (GLOBE NEWSWIRE) -- Today, AMD AMD announced the latest accelerator and networking solutions that will power the next generation of AI infrastructure at scale: AMD Instinct™ MI325X accelerators, the AMD Pensando™ Pollara 400 NIC and the AMD Pensando Salina DPU. AMD Instinct MI325X accelerators set a new standard in performance for Gen AI models and data centers. Built on the AMD CDNA™ 3 architecture, AMD Instinct MI325X accelerators are designed for exceptional performance and efficiency for demanding AI tasks spanning foundation model training, fine-tuning and inferencing. Together, these products enable AMD customers and partners to create highly performant and optimized AI solutions at the system, rack and data center level. "AMD continues to deliver on our roadmap, offering customers the performance they need and the choice they want, to bring AI infrastructure, at scale, to market faster," said Forrest Norrod, executive vice president and general manager, Data Center Solutions Business Group, AMD. "With the new AMD Instinct accelerators, EPYC processors and AMD Pensando networking engines, the continued growth of our open software ecosystem, and the ability to tie this all together into optimized AI infrastructure, AMD underscores the critical expertise to build and deploy world class AI solutions." AMD Instinct MI325X Extends Leading AI Performance AMD Instinct MI325X accelerators deliver industry-leading memory capacity and bandwidth, with 256GB of HBM3E supporting 6.0TB/s offering 1.8X more capacity and 1.3x more bandwidth than the H2001. The AMD Instinct MI325X also offers 1.3X greater peak theoretical FP16 and FP8 compute performance compared to H2001. This leadership memory and compute can provide up to 1.3X the inference performance on Mistral 7B at FP162, 1.2X the inference performance on Llama 3.1 70B at FP83 and 1.4X the inference performance on Mixtral 8x7B at FP16 of the H2004. AMD Instinct MI325X accelerators are currently on track for production shipments in Q4 2024 and are expected to have widespread system availability from a broad set of platform providers, including Dell Technologies, Eviden, Gigabyte, Hewlett Packard Enterprise, Lenovo, Supermicro and others starting in Q1 2025. Continuing its commitment to an annual roadmap cadence, AMD previewed the next-generation AMD Instinct MI350 series accelerators. Based on AMD CDNA 4 architecture, AMD Instinct MI350 series accelerators are designed to deliver a 35x improvement in inference performance compared to AMD CDNA 3-based accelerators5. The AMD Instinct MI350 series will continue to drive memory capacity leadership with up to 288GB of HBM3E memory per accelerator. The AMD Instinct MI350 series accelerators are on track to be available during the second half of 2025. AMD Next-Gen AI Networking AMD is leveraging the most widely deployed programmable DPU for hyperscalers to power next-gen AI networking. Split into two parts: the front-end, which delivers data and information to an AI cluster, and the backend, which manages data transfer between accelerators and clusters, AI networking is critical to ensuring CPUs and accelerators are utilized efficiently in AI infrastructure. To effectively manage these two networks and drive high performance, scalability and efficiency across the entire system, AMD introduced the AMD Pensando™ Salina DPU for the front-end and the AMD Pensando™ Pollara 400, the industry's first Ultra Ethernet Consortium (UEC) ready AI NIC, for the back-end. The AMD Pensando Salina DPU is the third generation of the world's most performant and programmable DPU, bringing up to 2X the performance, bandwidth and scale compared to the previous generation. Supporting 400G throughput for fast data transfer rates, the AMD Pensando Salina DPU is a critical component in AI front-end network clusters, optimizing performance, efficiency, security and scalability for data-driven AI applications. The UEC-ready AMD Pensando Pollara 400, powered by the AMD P4 Programmable engine, is the industry's first UEC-ready AI NIC. It supports the next-gen RDMA software and is backed by an open ecosystem of networking. The AMD Pensando Pollara 400 is critical for providing leadership performance, scalability and efficiency of accelerator-to-accelerator communication in back-end networks. Both the AMD Pensando Salina DPU and AMD Pensando Pollara 400 are sampling with customers in Q4'24 and are on track for availability in the first half of 2025. AMD AI Software Delivering New Capabilities for Generative AI AMD continues its investment in driving software capabilities and the open ecosystem to deliver powerful new features and capabilities in the AMD ROCm™ open software stack. Within the open software community, AMD is driving support for AMD compute engines in the most widely used AI frameworks, libraries and models including PyTorch, Triton, Hugging Face and many others. This work translates to out-of-the-box performance and support with AMD Instinct accelerators on popular generative AI models like Stable Diffusion 3, Meta Llama 3, 3.1 and 3.2 and more than one million models at Hugging Face. Beyond the community, AMD continues to advance its ROCm open software stack, bringing the latest features to support leading training and inference on Generative AI workloads. ROCm 6.2 now includes support for critical AI features like FP8 datatype, Flash Attention 3, Kernel Fusion and more. With these new additions, ROCm 6.2, compared to ROCm 6.0, provides up to a 2.4X performance improvement on inference6 and 1.8X on training for a variety of LLMs7. Supporting Resources Follow AMD on LinkedInFollow AMD on TwitterRead more about AMD Next Generation AI Networking hereRead more about AMD Instinct Accelerators hereVisit the AMD Advancing AI: 2024 event page About AMD For more than 50 years AMD has driven innovation in high-performance computing, graphics, and visualization technologies. Billions of people, leading Fortune 500 businesses, and cutting-edge scientific research institutions around the world rely on AMD technology daily to improve how they live, work, and play. AMD employees are focused on building leadership high-performance and adaptive products that push the boundaries of what is possible. For more information about how AMD is enabling today and inspiring tomorrow, visit the AMD AMD website, blog, LinkedIn, and X pages. CAUTIONARY STATEMENT This press release contains forward-looking statements concerning Advanced Micro Devices, Inc. (AMD) such as the features, functionality, performance, availability, timing and expected benefits of AMD products including the AMD Instinct™ MI325X accelerators; AMD Pensando™ Salina DPU; AMD Pensando Pollara 400; continued growth of AMD's open software ecosystem; AMD Instinct MI350 series accelerators, which are made pursuant to the Safe Harbor provisions of the Private Securities Litigation Reform Act of 1995. Forward-looking statements are commonly identified by words such as "would," "may," "expects," "believes," "plans," "intends," "projects" and other terms with similar meaning. Investors are cautioned that the forward-looking statements in this press release are based on current beliefs, assumptions and expectations, speak only as of the date of this press release and involve risks and uncertainties that could cause actual results to differ materially from current expectations. Such statements are subject to certain known and unknown risks and uncertainties, many of which are difficult to predict and generally beyond AMD's control, that could cause actual results and other future events to differ materially from those expressed in, or implied or projected by, the forward-looking information and statements. Material factors that could cause actual results to differ materially from current expectations include, without limitation, the following: Intel Corporation's dominance of the microprocessor market and its aggressive business practices; Nvidia's dominance in the graphics processing unit market and its aggressive business practices; the cyclical nature of the semiconductor industry; market conditions of the industries in which AMD products are sold; loss of a significant customer; competitive markets in which AMD's products are sold; economic and market uncertainty; quarterly and seasonal sales patterns; AMD's ability to adequately protect its technology or other intellectual property; unfavorable currency exchange rate fluctuations; ability of third party manufacturers to manufacture AMD's products on a timely basis in sufficient quantities and using competitive technologies; availability of essential equipment, materials, substrates or manufacturing processes; ability to achieve expected manufacturing yields for AMD's products; AMD's ability to introduce products on a timely basis with expected features and performance levels; AMD's ability to generate revenue from its semi-custom SoC products; potential security vulnerabilities; potential security incidents including IT outages, data loss, data breaches and cyberattacks; uncertainties involving the ordering and shipment of AMD's products; AMD's reliance on third-party intellectual property to design and introduce new products; AMD's reliance on third-party companies for design, manufacture and supply of motherboards, software, memory and other computer platform components; AMD's reliance on Microsoft and other software vendors' support to design and develop software to run on AMD's products; AMD's reliance on third-party distributors and add-in-board partners; impact of modification or interruption of AMD's internal business processes and information systems; compatibility of AMD's products with some or all industry-standard software and hardware; costs related to defective products; efficiency of AMD's supply chain; AMD's ability to rely on third party supply-chain logistics functions; AMD's ability to effectively control sales of its products on the gray market; long-term impact of climate change on AMD's business; impact of government actions and regulations such as export regulations, tariffs and trade protection measures; AMD's ability to realize its deferred tax assets; potential tax liabilities; current and future claims and litigation; impact of environmental laws, conflict minerals related provisions and other laws or regulations; evolving expectations from governments, investors, customers and other stakeholders regarding corporate responsibility matters; issues related to the responsible use of AI; restrictions imposed by agreements governing AMD's notes, the guarantees of Xilinx's notes and the revolving credit agreement; impact of acquisitions, joint ventures and/or investments on AMD's business and AMD's ability to integrate acquired businesses; impact of any impairment of the combined company's assets; political, legal and economic risks and natural disasters; future impairments of technology license purchases; AMD's ability to attract and retain qualified personnel; and AMD's stock price volatility. Investors are urged to review in detail the risks and uncertainties in AMD's Securities and Exchange Commission filings, including but not limited to AMD's most recent reports on Forms 10-K and 10-Q. AMD, the AMD Arrow logo, AMD CDNA, AMD Instinct, Pensando, ROCm, and combinations thereof are trademarks of Advanced Micro Devices, Inc. Other names are for informational purposes only and may be trademarks of their respective owners. ________________________________ 1MI325-002 -Calculations conducted by AMD Performance Labs as of May 28th, 2024 for the AMD Instinct™ MI325X GPU resulted in 1307.4 TFLOPS peak theoretical half precision (FP16), 1307.4 TFLOPS peak theoretical Bfloat16 format precision (BF16), 2614.9 TFLOPS peak theoretical 8-bit precision (FP8), 2614.9 TOPs INT8 floating-point performance. Actual performance will vary based on final specifications and system configuration. Published results on Nvidia H200 SXM (141GB) GPU: 989.4 TFLOPS peak theoretical half precision tensor (FP16 Tensor), 989.4 TFLOPS peak theoretical Bfloat16 tensor format precision (BF16 Tensor), 1,978.9 TFLOPS peak theoretical 8-bit precision (FP8), 1,978.9 TOPs peak theoretical INT8 floating-point performance. BFLOAT16 Tensor Core, FP16 Tensor Core, FP8 Tensor Core and INT8 Tensor Core performance were published by Nvidia using sparsity; for the purposes of comparison, AMD converted these numbers to non-sparsity/dense by dividing by 2, and these numbers appear above. Nvidia H200 source: https://nvdam.widen.net/s/nb5zzzsjdf/hpc-datasheet-sc23-h200-datasheet-3002446 and https://www.anandtech.com/show/21136/nvidia-at-sc23-h200-accelerator-with-hbm3e-and-jupiter-supercomputer-for-2024 Note: Nvidia H200 GPUs have the same published FLOPs performance as H100 products https://resources.nvidia.com/en-us-tensor-core/. 2 Based on testing completed on 9/28/2024 by AMD performance lab measuring overall latency for Mistral-7B model using FP16 datatype. Test was performed using input length of 128 tokens and an output length of 128 tokens for the following configurations of AMD Instinct™ MI325X GPU accelerator and NVIDIA H200 SXM GPU accelerator. 1x MI325X at 1000W with vLLM performance: 0.637 sec (latency in seconds) Vs. 1x H200 at 700W with TensorRT-LLM: 0.811 sec (latency in seconds) Configurations: AMD Instinct™ MI325X reference platform: 1x AMD Ryzen™ 9 7950X 16-Core Processor CPU, 1x AMD Instinct MI325X (256GiB, 1000W) GPU, Ubuntu® 22.04, and ROCm™ 6.3 pre-release Vs NVIDIA H200 HGX platform: Supermicro SuperServer with 2x Intel Xeon® Platinum 8468 Processors, 8x Nvidia H200 (140GB, 700W) GPUs [only 1 GPU was used in this test], Ubuntu 22.04), CUDA 12.6 Server manufacturers may vary configurations, yielding different results. Performance may vary based on use of latest drivers and optimizations. MI325-005 3 MI325-006: Based on testing completed on 9/28/2024 by AMD performance lab measuring overall latency for LLaMA 3.1-70B model using FP8 datatype. Test was performed using input length of 2048 tokens and an output length of 2048 tokens for the following configurations of AMD Instinct™ MI325X GPU accelerator and NVIDIA H200 SXM GPU accelerator. 1x MI325X at 1000W with vLLM performance: 48.025 sec (latency in seconds) Vs. 1x H200 at 700W with TensorRT-LLM: 62.688 sec (latency in seconds) Configurations: AMD Instinct™ MI325X reference platform: 1x AMD Ryzen™ 9 7950X 16-Core Processor CPU, 1x AMD Instinct MI325X (256GiB, 1000W) GPU, Ubuntu® 22.04, and ROCm™ 6.3 pre-release Vs NVIDIA H200 HGX platform: Supermicro SuperServer with 2x Intel Xeon® Platinum 8468 Processors, 8x Nvidia H200 (140GB, 700W) GPUs, Ubuntu 22.04), CUDA 12.6 Server manufacturers may vary configurations, yielding different results. Performance may vary based on use of latest drivers and optimizations. 4 MI325-004: Based on testing completed on 9/28/2024 by AMD performance lab measuring text generated throughput for Mixtral-8x7B model using FP16 datatype. Test was performed using input length of 128 tokens and an output length of 4096 tokens for the following configurations of AMD Instinct™ MI325X GPU accelerator and NVIDIA H200 SXM GPU accelerator. 1x MI325X at 1000W with vLLM performance: 4598 (Output tokens / sec) Vs. 1x H200 at 700W with TensorRT-LLM: 2700.7 (Output tokens / sec) Configurations: AMD Instinct™ MI325X reference platform: 1x AMD Ryzen™ 9 7950X CPU, 1x AMD Instinct MI325X (256GiB, 1000W) GPU, Ubuntu® 22.04, and ROCm™ 6.3 pre-release Vs NVIDIA H200 HGX platform: Supermicro SuperServer with 2x Intel Xeon® Platinum 8468 Processors, 8x Nvidia H200 (140GB, 700W) GPUs [only 1 GPU was used in this test], Ubuntu 22.04) CUDA® 12.6 Server manufacturers may vary configurations, yielding different results. Performance may vary based on use of latest drivers and optimizations. 5 CDNA4-03: Inference performance projections as of May 31, 2024 using engineering estimates based on the design of a future AMD CDNA 4-based Instinct MI350 Series accelerator as proxy for projected AMD CDNA™ 4 performance. A 1.8T GPT MoE model was evaluated assuming a token-to-token latency = 70ms real time, first token latency = 5s, input sequence length = 8k, output sequence length = 256, assuming a 4x 8-mode MI350 series proxy (CDNA4) vs. 8x MI300X per GPU performance comparison.. Actual performance will vary based on factors including but not limited to final specifications of production silicon, system configuration and inference model and size used. 6 MI300-62: Testing conducted by internal AMD Performance Labs as of September 29, 2024 inference performance comparison between ROCm 6.2 software and ROCm 6.0 software on the systems with 8 AMD Instinct™ MI300X GPUs coupled with Llama 3.1-8B, Llama 3.1-70B, Mixtral-8x7B, Mixtral-8x22B, and Qwen 72B models. ROCm 6.2 with vLLM 0.5.5 performance was measured against the performance with ROCm 6.0 with vLLM 0.3.3, and tests were performed across batch sizes of 1 to 256 and sequence lengths of 128 to 2048. Configurations: 1P AMD EPYC™ 9534 CPU server with 8x AMD Instinct™ MI300X (192GB, 750W) GPUs, Supermicro AS-8125GS-TNMR2, NPS1 (1 NUMA per socket), 1.5 TiB (24 DIMMs, 4800 mts memory, 64 GiB/DIMM), 4x 3.49TB Micron 7450 storage, BIOS version: 1.8, , ROCm 6.2.0-00, vLLM 0.5.5, PyTorch 2.4.0, Ubuntu® 22.04 LTS with Linux kernel 5.15.0-119-generic. vs. 1P AMD EPYC 9534 CPU server with 8x AMD Instinct™ MI300X (192GB, 750W) GPUs, Supermicro AS-8125GS-TNMR2, NPS1 (1 NUMA per socket), 1.5TiB 24 DIMMs, 4800 mts memory, 64 GiB/DIMM), 4x 3.49TB Micron 7450 storage, BIOS version: 1.8, ROCm 6.0.0-00, vLLM 0.3.3, PyTorch 2.1.1, Ubuntu 22.04 LTS with Linux kernel 5.15.0-119-generic. Server manufacturers may vary configurations, yielding different results. Performance may vary based on factors including but not limited to different versions of configurations, vLLM, and drivers. 7 MI300-61: Measurements conducted by AMD AI Product Management team on AMD Instinct™ MI300X GPU for comparing large language model (LLM) performance with optimization methodologies enabled and disabled as of 9/28/2024 on Llama 3.1-70B and Llama 3.1-405B and vLLM 0.5.5. System Configurations: - AMD EPYC 9654 96-Core Processor, 8 x AMD MI300X, ROCm™ 6.1, Linux® 7ee7e017abe3 5.15.0-116-generic #126-Ubuntu® SMP Mon Jul 1 10:14:24 UTC 2024 x86_64 x86_64 x86_64 GNU/Linux, Frequency boost: enabled. Performance may vary on factors including but not limited to different versions of configurations, vLLM, and drivers. Aaron Grabein AMD Communications +1 737-256-9518 aaron.grabein@amd.com Mitch Haws AMD Investor Relations +1 512-944-0790 mitch.haws@amd.com Market News and Data brought to you by Benzinga APIs
[15]
AMD Launches New Ryzen™ AI PRO 300 Series Processors to Power Next Generation of Commercial PCs - Advanced Micro Devices (NASDAQ:AMD)
- New processors deliver unprecedented AI compute capabilities1 and multi-day battery life2, enabling incredible productivity for business users - - AMD continues to expand commercial portfolio; more than 100 Ryzen AI PRO PCs on-track to launch through 2025 - SAN FRANCISCO, Oct. 10, 2024 (GLOBE NEWSWIRE) -- Today, AMD AMD announced its third generation commercial AI mobile processors, designed specifically to transform business productivity with Copilot+ features including live captioning and language translation in conference calls and advanced AI image generators. The new Ryzen AI PRO 300 Series processors deliver industry-leading AI compute3, with up to three times the AI performance than the previous generation4, and offer uncompromising performance for everyday workloads. Enabled with AMD PRO Technologies, the Ryzen AI PRO 300 Series processors offer world-class security and manageability features designed to streamline IT operations and ensure exceptional ROI for businesses. Ryzen AI PRO 300 Series processors feature new AMD "Zen 5" architecture, delivering outstanding CPU performance, and are the world's best line up of commercial processors for Copilot+ enterprise PCs5. Laptops equipped with Ryzen AI PRO 300 Series processors are designed to tackle business' toughest workloads, with the top-of-stack Ryzen AI 9 HX PRO 375 offering up to 40% higher performance6 and up to 14% faster productivity performance7 compared to Intel's Core Ultra 7 165U. With the addition of XDNA™ 2 architecture powering the integrated NPU, AMD Ryzen AI PRO 300 Series processors offer a cutting-edge 50+ NPU TOPS (Trillions of Operations Per Second) of AI processing power, exceeding Microsoft's Copilot+ AI PC requirements89 and delivering exceptional AI compute and productivity capabilities for the modern business. Built on a 4nm process and with innovative power management, the new processors deliver extended battery life ideal for sustained performance and productivity on the go. "Enterprises are increasingly demanding more compute power and efficiency to drive their everyday tasks and most taxing workloads. We are excited to add the Ryzen AI PRO 300 Series, the most powerful AI processor built for business PCs10, to our portfolio of mobile processors," said Jack Huynh, senior vice president and general manager, Computing and Graphics Group at AMD. "Our third generation AI-enabled processors for business PCs deliver unprecedented AI processing capabilities with incredible battery life and seamless compatibility for the applications users depend on." AMD Ryzen AI PRO 300 Series Mobile Processors ModelCores/ThreadsBoost11 / Base FrequencyTotal CacheGraphics Model AMD cTDPTOPSAMD Ryzen™ AI 9 HX PRO 37512C/24TUp to 5.1GHz/ 2GHz36MBRadeon™ 890M Graphics15-54WUp to 55AMD Ryzen™ AI 9 HX PRO 37012C/24TUp to 5.1GHz/ 2GHz36MBRadeon™ 890M Graphics15-54WUp to 50AMD Ryzen™ AI 7 PRO 3608C/16TUp to 5GHz/ 2GHz24MBAMD Radeon™ 880M Graphics15-54WUp to 50 AMD Continues to Expand Commercial OEM Ecosystem OEM partners continue to expand their commercial offerings with new PCs powered by Ryzen AI PRO 300 Series processors, delivering well-rounded performance and compatibility to their business customers. With industry leading TOPS, the next generation of Ryzen processor-powered commercial PCs are set to expand the possibilities of local AI processing with Microsoft Copilot+. OEM systems powered by Ryzen AI PRO 300 Series are expected to be on shelf starting later this year. "Microsoft's partnership with AMD and the integration of Ryzen AI PRO processors into Copilot+ PCs demonstrate our joint focus on delivering impactful AI-driven experiences for our customers. The Ryzen AI PRO's performance, combined with the latest features in Windows 11, enhances productivity, efficiency, and security," said Pavan Davuluri, corporate vice president, Windows+ Devices, Microsoft. "Features like Improved Windows Search, Recall, and Click to Do make PCs more intuitive and responsive. Security enhancements, including the Microsoft Pluton security processor and Windows Hello Enhanced Sign-in Security, help safeguard customer data with advanced protection. We're proud of our strong history of collaboration with AMD and are thrilled to bring these innovations to market." "In today's AI-powered era of computing, HP is dedicated to delivering powerful innovation and performance that revolutionizes the way people work," said Alex Cho, president of Personal Systems, HP. "With the HP EliteBook X Next-Gen AI PC, we are empowering modern leaders to push boundaries without compromising power or performance. We are proud to expand our AI PC lineup powered by AMD, providing our commercial customers with a truly personalized experience." "Lenovo's partnership with AMD continues to drive AI PC innovation and deliver supreme performance for our business customers. Our recently announced ThinkPad T14s Gen 6 AMD, powered by the latest AMD Ryzen AI PRO 300 Series processors, showcases the strength of our collaboration," said Luca Rossi, president, Lenovo Intelligent Devices Group. "This device offers outstanding AI computing power, enhanced security, and exceptional battery life, providing professionals with the tools they need to maximize productivity and efficiency. Together with AMD, we are transforming the business landscape by delivering smarter, AI-driven solutions that empower users to achieve more." New PRO Technologies Features Build Upon Leadership Security and Management Features In addition to AMD Secure Processor12, AMD Shadow Stack and AMD Platform Secure Boot, AMD has expanded its PRO Technologies lineup with new security and manageability features. Processors equipped with PRO Technologies will now come standard with Cloud Bare Metal Recovery, allowing IT teams to seamlessly recover systems via the cloud ensuring smooth and continuous operations; Supply Chain Security (AMD Device Identity), a new supply chain security function, enabling traceability across the supply chain; and Watch Dog Timer, building on existing resiliency support with additional detection and recovery processes. Additional AI-based malware detection is available via PRO Technologies with select ISV partners. These new security features leverage the integrated NPU to run AI-based security workloads without impacting day-to-day performance. Supporting Resources Learn more about Ryzen PRO mobile processorsLearn more about AMD PRO TechnologiesLearn more about Ryzen AILearn more about Ryzen AI SoftwareBecome a fan of AMD on FacebookFollow AMD on X About AMD For more than 50 years AMD has driven innovation in high-performance computing, graphics and visualization technologies. Billions of people, leading Fortune 500 businesses and cutting-edge scientific research institutions around the world rely on AMD technology daily to improve how they live, work and play. AMD employees are focused on building leadership high-performance and adaptive products that push the boundaries of what is possible. For more information about how AMD is enabling today and inspiring tomorrow, visit the AMD AMD website, blog, LinkedIn and X pages. Cautionary Statement This press release contains forward-looking statements concerning Advanced Micro Devices, Inc. (AMD) such as the features, functionality, performance, availability, timing and expected benefits of AMD products including the AMD Ryzen™ AI PRO 300 Series mobile processors, which are made pursuant to the Safe Harbor provisions of the Private Securities Litigation Reform Act of 1995. Forward-looking statements are commonly identified by words such as "would," "may," "expects," "believes," "plans," "intends," "projects" and other terms with similar meaning. Investors are cautioned that the forward-looking statements in this press release are based on current beliefs, assumptions and expectations, speak only as of the date of this press release and involve risks and uncertainties that could cause actual results to differ materially from current expectations. Such statements are subject to certain known and unknown risks and uncertainties, many of which are difficult to predict and generally beyond AMD's control, that could cause actual results and other future events to differ materially from those expressed in, or implied or projected by, the forward-looking information and statements. Material factors that could cause actual results to differ materially from current expectations include, without limitation, the following: Intel Corporation's dominance of the microprocessor market and its aggressive business practices; Nvidia's dominance in the graphics processing unit market and its aggressive business practices; the cyclical nature of the semiconductor industry; market conditions of the industries in which AMD products are sold; loss of a significant customer; competitive markets in which AMD's products are sold; economic and market uncertainty; quarterly and seasonal sales patterns; AMD's ability to adequately protect its technology or other intellectual property; unfavorable currency exchange rate fluctuations; ability of third party manufacturers to manufacture AMD's products on a timely basis in sufficient quantities and using competitive technologies; availability of essential equipment, materials, substrates or manufacturing processes; ability to achieve expected manufacturing yields for AMD's products; AMD's ability to introduce products on a timely basis with expected features and performance levels; AMD's ability to generate revenue from its semi-custom SoC products; potential security vulnerabilities; potential security incidents including IT outages, data loss, data breaches and cyberattacks; uncertainties involving the ordering and shipment of AMD's products; AMD's reliance on third-party intellectual property to design and introduce new products; AMD's reliance on third-party companies for design, manufacture and supply of motherboards, software, memory and other computer platform components; AMD's reliance on Microsoft and other software vendors' support to design and develop software to run on AMD's products; AMD's reliance on third-party distributors and add-in-board partners; impact of modification or interruption of AMD's internal business processes and information systems; compatibility of AMD's products with some or all industry-standard software and hardware; costs related to defective products; efficiency of AMD's supply chain; AMD's ability to rely on third party supply-chain logistics functions; AMD's ability to effectively control sales of its products on the gray market; long-term impact of climate change on AMD's business; impact of government actions and regulations such as export regulations, tariffs and trade protection measures; AMD's ability to realize its deferred tax assets; potential tax liabilities; current and future claims and litigation; impact of environmental laws, conflict minerals related provisions and other laws or regulations; evolving expectations from governments, investors, customers and other stakeholders regarding corporate responsibility matters; issues related to the responsible use of AI; restrictions imposed by agreements governing AMD's notes, the guarantees of Xilinx's notes and the revolving credit agreement; impact of acquisitions, joint ventures and/or investments on AMD's business and AMD's ability to integrate acquired businesses; impact of any impairment of the combined company's assets; political, legal and economic risks and natural disasters; future impairments of technology license purchases; AMD's ability to attract and retain qualified personnel; and AMD's stock price volatility. Investors are urged to review in detail the risks and uncertainties in AMD's Securities and Exchange Commission filings, including but not limited to AMD's most recent reports on Forms 10-K and 10-Q. © 2024 Advanced Micro Devices, Inc. All rights reserved. AMD, the AMD Arrow logo, Radeon, RDNA, Ryzen, XDNA and combinations thereof are trademarks of Advanced Micro Devices, Inc. Certain AMD technologies may require third-party enablement or activation. Supported features may vary by operating system. Please confirm with the system manufacturer for specific features. No technology or product can be completely secure. The information contained herein is for informational purposes only and is subject to change without notice. Timelines, roadmaps, and/or product release dates shown in this Press Release are plans only and subject to change. 1 As of May 2023, AMD has the first available dedicated AI engine on an x86 Windows processor, where 'dedicated AI engine' is defined as an AI engine that has no function other than to process AI inference models and is part of the x86 processor die. For detailed information, please check: https://www.amd.com/en/technologies/xdna.html. PHX-3a. 2 All battery life claims are approximate. Actual battery life will vary based on several factors, including, but not limited to: product configuration and usage, software, operating conditions, wireless functionality, power management settings, screen brightness and other factors. The maximum capacity of the battery will naturally decrease with time and use. AMD has not independently tested or verified the battery life claim. GD-168. 3 Based on AMD product specifications and competitive products announced as of Oct 2024. AMD Ryzen™ AI PRO 300 Series processors' NPU offers up to 55 peak TOPS. This is the most TOPS offered on any system found in enterprise today. AI PC is defined as a laptop PC with a processor that includes a neural processing unit (NPU). STXP-06. 4 Based on TOPS specification of AMD Ryzen™ AI 300 Series processors with 50 TOPS compared to an AMD Ryzen 8040 Series processors with 16 TOPS as of June 2024. STX-01. 5 Based on product specifications and competitive products announced as of Oct 2024 and testing as of Sept 2024 by AMD performance labs using the following systems: HP EliteBook X G1a with AMD Ryzen AI 9 HX PRO 375 processor @23W, Radeon 880M graphics, 32GB of RAM, 512GB SSD, VBS=ON, Windows 11 PRO; Dell Latitude 7450 with Intel Core Ultra 7 165U processor @15W (vPro enabled), Intel Iris Xe Graphics, VBS=ON, 32GB RAM, 512GB NVMe SSD, Microsoft Windows 11 Professional; Dell Latitude 7450 with Intel Core Ultra 7 165H processor @28W (vPro enabled), Intel Iris Xe Graphics, VBS=ON, 16GB RAM, 512GB NVMe SSD, Microsoft Windows 11 Pro. All systems were tested in Best Performance Mode. AI PC is defined as a laptop PC with a processor that includes a neural processing unit (NPU). STXP-04. 6 Testing as of Sept 2024 by AMD performance labs on an HP EliteBook X G1a (14in) (40W) with AMD Ryzen AI 9 HX PRO 375 processor, Radeon™ 890M graphics, 32GB of RAM, 512GB SSD, VBS=ON, Windows 11 Pro vs. a Dell Latitude 7450 with an Intel Core Ultra 7 165H processor (vPro enabled), Intel Arc Graphics, VBS=ON, 16GB RAM, 512GB NVMe SSD, Microsoft Windows 11 Pro in the application(s) (Best Performance Mode): Cinebench R24 nT. Laptop manufactures may vary configurations yielding different results. STXP-12. 7 Testing as of Sept 2024 by AMD performance labs using the following systems: (1) HP EliteBook X G1a with AMD Ryzen AI 9 HX PRO 375 processor (@40W), Radeon™ 890M graphics, 32GB of RAM, 512GB SSD, VBS=ON, Windows 11 Pro; (2) Dell Latitude 7450 with Intel Core Ultra 7 165U processor (@15W) (vPro enabled), Intel Iris Xe Graphics, VBS=ON, 32GB RAM, 512GB NVMe SSD, Microsoft Windows 11 Professional; and (3) Dell Latitude 7450 with Intel Core Ultra 7 165H processor (@28W) (vPro enabled), Intel Integrated, VBS=ON, 16GB RAM, 512GB NVMe SSD, Microsoft Windows 11 Pro. Tested applications (in Balanced Mode) include: Procyon Office Productivity, Procyon Office Productivity Excel, Procyon Office Productivity Outlook, Procyon Office Productivity Power Point, Procyon Office Productivity Word, Composite Geomean Score. Laptop manufactures may vary configurations yielding different results. STXP-18. 8 Based on Microsoft Copilot+ requirements of minimum 40 TOPS using AMD product specifications and competitive products announced as of Oct 2024. Microsoft requirements found here - https://support.microsoft.com/en-us/topic/copilot-pc-hardware-requirements-35782169-6eab-4d63-a5c5-c498c3037364. STXP-05. 9 Trillions of Operations per Second (TOPS) for an AMD Ryzen processor is the maximum number of operations per second that can be executed in an optimal scenario and may not be typical. TOPS may vary based on several factors, including the specific system configuration, AI model, and software version. GD-243. 10 Testing as of Sept 2024 by AMD performance labs using the following benchmarks: Blender, Cinebench R24, Geekbench 6.3, and Passmark 11, systems: HP EliteBook X G1a with AMD Ryzen AI 9 HX PRO 375 processor @54W, Radeon 880M graphics, 32GB of RAM, 512GB SSD; Lenovo ThinkPad T14s Gen 6 with AMD Ryzen™ AI 7 PRO 360 processor @22W, Radeon™ 880M graphics, 32GB RAM, 1TB SSD; Dell Latitude 7450 with Intel Core Ultra 7 165U processor @15W (vPro enabled), Intel Iris Xe Graphics, 32GB RAM, 512GB NVMe SSD; Dell Latitude 7450 with Intel Core Ultra 7 165H processor @28W (vPro enabled), Intel Iris Xe Graphics, 16GB RAM, 512GB NVMe SSD,. All systems Windows 11 Pro, VBS=ON, and tested in Best Performance Mode. PassMark is a registered trademark of PassMark Software Pty Ltd. AI PC is defined as a laptop PC with a processor that includes a neural processing unit (NPU). STXP-07. 11 Boost Clock Frequency is the maximum frequency achievable on the CPU running a bursty workload. Boost clock achievability, frequency, and sustainability will vary based on several factors, including but not limited to: thermal conditions and variation in applications and workloads. GD-150 12 The AMD Secure Processor is a dedicated on-chip security processor integrated within each system-on-a-chip (SoC) and ASIC (Application Specific Integrated Circuit) designed by AMD. It enables secure boot with root of trust anchored in hardware, initializes the SoC through a secure boot flow, and establishes an isolated Trusted Execution Environment. GD-72. Stacy MacDiarmid AMD Communications +1 512-658-2265 Stacy.MacDiarmid@amd.com Mitch Haws AMD Investor Relations +1 512-944-0790 Mitch.Haws@amd.com A photo accompanying this announcement is available at https://www.globenewswire.com/NewsRoom/AttachmentNg/c67477ae-0d96-4936-91ba-cd836bfa321e Market News and Data brought to you by Benzinga APIs
[16]
AMD Launches New Ryzen™ AI PRO 300 Series Processors to Power Next Generation of Commercial PCs
- New processors deliver unprecedented AI compute capabilities and multi-day battery life, enabling incredible productivity for business users - - AMD continues to expand commercial portfolio; more than 100 Ryzen AI PRO PCs on-track to launch through 2025 - SAN FRANCISCO, Oct. 10, 2024 (GLOBE NEWSWIRE) -- Today, AMD (NASDAQ: AMD) announced its third generation commercial AI mobile processors, designed specifically to transform business productivity with Copilot+ features including live captioning and language translation in conference calls and advanced AI image generators. The new Ryzen AI PRO 300 Series processors deliver industry-leading AI compute, with up to three times the AI performance than the previous generation, and offer uncompromising performance for everyday workloads. Enabled with AMD PRO Technologies, the Ryzen AI PRO 300 Series processors offer world-class security and manageability features designed to streamline IT operations and ensure exceptional ROI for businesses. Ryzen AI PRO 300 Series processors feature new AMD "Zen 5" architecture, delivering outstanding CPU performance, and are the world's best line up of commercial processors for Copilot+ enterprise PCs. Laptops equipped with Ryzen AI PRO 300 Series processors are designed to tackle business' toughest workloads, with the top-of-stack Ryzen AI 9 HX PRO 375 offering up to 40% higher performance and up to 14% faster productivity performance compared to Intel's Core Ultra 7 165U. With the addition of XDNA™ 2 architecture powering the integrated NPU, AMD Ryzen AI PRO 300 Series processors offer a cutting-edge 50+ NPU TOPS (Trillions of Operations Per Second) of AI processing power, exceeding Microsoft's Copilot+ AI PC requirements and delivering exceptional AI compute and productivity capabilities for the modern business. Built on a 4nm process and with innovative power management, the new processors deliver extended battery life ideal for sustained performance and productivity on the go. "Enterprises are increasingly demanding more compute power and efficiency to drive their everyday tasks and most taxing workloads. We are excited to add the Ryzen AI PRO 300 Series, the most powerful AI processor built for business PCs, to our portfolio of mobile processors," said Jack Huynh, senior vice president and general manager, Computing and Graphics Group at AMD. "Our third generation AI-enabled processors for business PCs deliver unprecedented AI processing capabilities with incredible battery life and seamless compatibility for the applications users depend on." AMD Ryzen AI PRO 300 Series Mobile Processors AMD Continues to Expand Commercial OEM Ecosystem OEM partners continue to expand their commercial offerings with new PCs powered by Ryzen AI PRO 300 Series processors, delivering well-rounded performance and compatibility to their business customers. With industry leading TOPS, the next generation of Ryzen processor-powered commercial PCs are set to expand the possibilities of local AI processing with Microsoft Copilot+. OEM systems powered by Ryzen AI PRO 300 Series are expected to be on shelf starting later this year. "Microsoft's partnership with AMD and the integration of Ryzen AI PRO processors into Copilot+ PCs demonstrate our joint focus on delivering impactful AI-driven experiences for our customers. The Ryzen AI PRO's performance, combined with the latest features in Windows 11, enhances productivity, efficiency, and security," said Pavan Davuluri, corporate vice president, Windows+ Devices, Microsoft. "Features like Improved Windows Search, Recall, and Click to Do make PCs more intuitive and responsive. Security enhancements, including the Microsoft Pluton security processor and Windows Hello Enhanced Sign-in Security, help safeguard customer data with advanced protection. We're proud of our strong history of collaboration with AMD and are thrilled to bring these innovations to market." "In today's AI-powered era of computing, HP is dedicated to delivering powerful innovation and performance that revolutionizes the way people work," said Alex Cho, president of Personal Systems, HP. "With the HP EliteBook X Next-Gen AI PC, we are empowering modern leaders to push boundaries without compromising power or performance. We are proud to expand our AI PC lineup powered by AMD, providing our commercial customers with a truly personalized experience." "Lenovo's partnership with AMD continues to drive AI PC innovation and deliver supreme performance for our business customers. Our recently announced ThinkPad T14s Gen 6 AMD, powered by the latest AMD Ryzen AI PRO 300 Series processors, showcases the strength of our collaboration," said Luca Rossi, president, Lenovo Intelligent Devices Group. "This device offers outstanding AI computing power, enhanced security, and exceptional battery life, providing professionals with the tools they need to maximize productivity and efficiency. Together with AMD, we are transforming the business landscape by delivering smarter, AI-driven solutions that empower users to achieve more." New PRO Technologies Features Build Upon Leadership Security and Management Features In addition to AMD Secure Processor, AMD Shadow Stack and AMD Platform Secure Boot, AMD has expanded its PRO Technologies lineup with new security and manageability features. Processors equipped with PRO Technologies will now come standard with Cloud Bare Metal Recovery, allowing IT teams to seamlessly recover systems via the cloud ensuring smooth and continuous operations; Supply Chain Security (AMD Device Identity), a new supply chain security function, enabling traceability across the supply chain; and Watch Dog Timer, building on existing resiliency support with additional detection and recovery processes. Additional AI-based malware detection is available via PRO Technologies with select ISV partners. These new security features leverage the integrated NPU to run AI-based security workloads without impacting day-to-day performance. Supporting Resources About AMD For more than 50 years AMD has driven innovation in high-performance computing, graphics and visualization technologies. Billions of people, leading Fortune 500 businesses and cutting-edge scientific research institutions around the world rely on AMD technology daily to improve how they live, work and play. AMD employees are focused on building leadership high-performance and adaptive products that push the boundaries of what is possible. For more information about how AMD is enabling today and inspiring tomorrow, visit the AMD (NASDAQ: AMD) website, blog, LinkedIn and X pages. Cautionary Statement This press release contains forward-looking statements concerning Advanced Micro Devices, Inc. (AMD) such as the features, functionality, performance, availability, timing and expected benefits of AMD products including the AMD Ryzen™ AI PRO 300 Series mobile processors, which are made pursuant to the Safe Harbor provisions of the Private Securities Litigation Reform Act of 1995. Forward-looking statements are commonly identified by words such as "would," "may," "expects," "believes," "plans," "intends," "projects" and other terms with similar meaning. Investors are cautioned that the forward-looking statements in this press release are based on current beliefs, assumptions and expectations, speak only as of the date of this press release and involve risks and uncertainties that could cause actual results to differ materially from current expectations. Such statements are subject to certain known and unknown risks and uncertainties, many of which are difficult to predict and generally beyond AMD's control, that could cause actual results and other future events to differ materially from those expressed in, or implied or projected by, the forward-looking information and statements. Material factors that could cause actual results to differ materially from current expectations include, without limitation, the following: Intel Corporation's dominance of the microprocessor market and its aggressive business practices; Nvidia's dominance in the graphics processing unit market and its aggressive business practices; the cyclical nature of the semiconductor industry; market conditions of the industries in which AMD products are sold; loss of a significant customer; competitive markets in which AMD's products are sold; economic and market uncertainty; quarterly and seasonal sales patterns; AMD's ability to adequately protect its technology or other intellectual property; unfavorable currency exchange rate fluctuations; ability of third party manufacturers to manufacture AMD's products on a timely basis in sufficient quantities and using competitive technologies; availability of essential equipment, materials, substrates or manufacturing processes; ability to achieve expected manufacturing yields for AMD's products; AMD's ability to introduce products on a timely basis with expected features and performance levels; AMD's ability to generate revenue from its semi-custom SoC products; potential security vulnerabilities; potential security incidents including IT outages, data loss, data breaches and cyberattacks; uncertainties involving the ordering and shipment of AMD's products; AMD's reliance on third-party intellectual property to design and introduce new products; AMD's reliance on third-party companies for design, manufacture and supply of motherboards, software, memory and other computer platform components; AMD's reliance on Microsoft and other software vendors' support to design and develop software to run on AMD's products; AMD's reliance on third-party distributors and add-in-board partners; impact of modification or interruption of AMD's internal business processes and information systems; compatibility of AMD's products with some or all industry-standard software and hardware; costs related to defective products; efficiency of AMD's supply chain; AMD's ability to rely on third party supply-chain logistics functions; AMD's ability to effectively control sales of its products on the gray market; long-term impact of climate change on AMD's business; impact of government actions and regulations such as export regulations, tariffs and trade protection measures; AMD's ability to realize its deferred tax assets; potential tax liabilities; current and future claims and litigation; impact of environmental laws, conflict minerals related provisions and other laws or regulations; evolving expectations from governments, investors, customers and other stakeholders regarding corporate responsibility matters; issues related to the responsible use of AI; restrictions imposed by agreements governing AMD's notes, the guarantees of Xilinx's notes and the revolving credit agreement; impact of acquisitions, joint ventures and/or investments on AMD's business and AMD's ability to integrate acquired businesses; impact of any impairment of the combined company's assets; political, legal and economic risks and natural disasters; future impairments of technology license purchases; AMD's ability to attract and retain qualified personnel; and AMD's stock price volatility. Investors are urged to review in detail the risks and uncertainties in AMD's Securities and Exchange Commission filings, including but not limited to AMD's most recent reports on Forms 10-K and 10-Q. © 2024 Advanced Micro Devices, Inc. All rights reserved. AMD, the AMD Arrow logo, Radeon, RDNA, Ryzen, XDNA and combinations thereof are trademarks of Advanced Micro Devices, Inc. Certain AMD technologies may require third-party enablement or activation. Supported features may vary by operating system. Please confirm with the system manufacturer for specific features. No technology or product can be completely secure. The information contained herein is for informational purposes only and is subject to change without notice. Timelines, roadmaps, and/or product release dates shown in this Press Release are plans only and subject to change. As of May 2023, AMD has the first available dedicated AI engine on an x86 Windows processor, where 'dedicated AI engine' is defined as an AI engine that has no function other than to process AI inference models and is part of the x86 processor die. For detailed information, please check: https://www.amd.com/en/technologies/xdna.html. PHX-3a. All battery life claims are approximate. Actual battery life will vary based on several factors, including, but not limited to: product configuration and usage, software, operating conditions, wireless functionality, power management settings, screen brightness and other factors. The maximum capacity of the battery will naturally decrease with time and use. AMD has not independently tested or verified the battery life claim. GD-168. Based on AMD product specifications and competitive products announced as of Oct 2024. AMD Ryzen™ AI PRO 300 Series processors' NPU offers up to 55 peak TOPS. This is the most TOPS offered on any system found in enterprise today. AI PC is defined as a laptop PC with a processor that includes a neural processing unit (NPU). STXP-06. Based on TOPS specification of AMD Ryzen™ AI 300 Series processors with 50 TOPS compared to an AMD Ryzen 8040 Series processors with 16 TOPS as of June 2024. STX-01. Based on product specifications and competitive products announced as of Oct 2024 and testing as of Sept 2024 by AMD performance labs using the following systems: HP EliteBook X G1a with AMD Ryzen AI 9 HX PRO 375 processor @23W, Radeon 880M graphics, 32GB of RAM, 512GB SSD, VBS=ON, Windows 11 PRO; Dell Latitude 7450 with Intel Core Ultra 7 165U processor @15W (vPro enabled), Intel Iris Xe Graphics, VBS=ON, 32GB RAM, 512GB NVMe SSD, Microsoft Windows 11 Professional; Dell Latitude 7450 with Intel Core Ultra 7 165H processor @28W (vPro enabled), Intel Iris Xe Graphics, VBS=ON, 16GB RAM, 512GB NVMe SSD, Microsoft Windows 11 Pro. All systems were tested in Best Performance Mode. AI PC is defined as a laptop PC with a processor that includes a neural processing unit (NPU). STXP-04. Testing as of Sept 2024 by AMD performance labs on an HP EliteBook X G1a (14in) (40W) with AMD Ryzen AI 9 HX PRO 375 processor, Radeon™ 890M graphics, 32GB of RAM, 512GB SSD, VBS=ON, Windows 11 Pro vs. a Dell Latitude 7450 with an Intel Core Ultra 7 165H processor (vPro enabled), Intel Arc Graphics, VBS=ON, 16GB RAM, 512GB NVMe SSD, Microsoft Windows 11 Pro in the application(s) (Best Performance Mode): Cinebench R24 nT. Laptop manufactures may vary configurations yielding different results. STXP-12. Testing as of Sept 2024 by AMD performance labs using the following systems: (1) HP EliteBook X G1a with AMD Ryzen AI 9 HX PRO 375 processor (@40W), Radeon™ 890M graphics, 32GB of RAM, 512GB SSD, VBS=ON, Windows 11 Pro; (2) Dell Latitude 7450 with Intel Core Ultra 7 165U processor (@15W) (vPro enabled), Intel Iris Xe Graphics, VBS=ON, 32GB RAM, 512GB NVMe SSD, Microsoft Windows 11 Professional; and (3) Dell Latitude 7450 with Intel Core Ultra 7 165H processor (@28W) (vPro enabled), Intel Integrated, VBS=ON, 16GB RAM, 512GB NVMe SSD, Microsoft Windows 11 Pro. Tested applications (in Balanced Mode) include: Procyon Office Productivity, Procyon Office Productivity Excel, Procyon Office Productivity Outlook, Procyon Office Productivity Power Point, Procyon Office Productivity Word, Composite Geomean Score. Laptop manufactures may vary configurations yielding different results. STXP-18. Based on Microsoft Copilot+ requirements of minimum 40 TOPS using AMD product specifications and competitive products announced as of Oct 2024. Microsoft requirements found here - https://support.microsoft.com/en-us/topic/copilot-pc-hardware-requirements-35782169-6eab-4d63-a5c5-c498c3037364. STXP-05. Trillions of Operations per Second (TOPS) for an AMD Ryzen processor is the maximum number of operations per second that can be executed in an optimal scenario and may not be typical. TOPS may vary based on several factors, including the specific system configuration, AI model, and software version. GD-243. Testing as of Sept 2024 by AMD performance labs using the following benchmarks: Blender, Cinebench R24, Geekbench 6.3, and Passmark 11, systems: HP EliteBook X G1a with AMD Ryzen AI 9 HX PRO 375 processor @54W, Radeon 880M graphics, 32GB of RAM, 512GB SSD; Lenovo ThinkPad T14s Gen 6 with AMD Ryzen™ AI 7 PRO 360 processor @22W, Radeon™ 880M graphics, 32GB RAM, 1TB SSD; Dell Latitude 7450 with Intel Core Ultra 7 165U processor @15W (vPro enabled), Intel Iris Xe Graphics, 32GB RAM, 512GB NVMe SSD; Dell Latitude 7450 with Intel Core Ultra 7 165H processor @28W (vPro enabled), Intel Iris Xe Graphics, 16GB RAM, 512GB NVMe SSD,. All systems Windows 11 Pro, VBS=ON, and tested in Best Performance Mode. PassMark is a registered trademark of PassMark Software Pty Ltd. AI PC is defined as a laptop PC with a processor that includes a neural processing unit (NPU). STXP-07. Boost Clock Frequency is the maximum frequency achievable on the CPU running a bursty workload. Boost clock achievability, frequency, and sustainability will vary based on several factors, including but not limited to: thermal conditions and variation in applications and workloads. GD-150 The AMD Secure Processor is a dedicated on-chip security processor integrated within each system-on-a-chip (SoC) and ASIC (Application Specific Integrated Circuit) designed by AMD. It enables secure boot with root of trust anchored in hardware, initializes the SoC through a secure boot flow, and establishes an isolated Trusted Execution Environment. GD-72. A photo accompanying this announcement is available at https://www.globenewswire.com/NewsRoom/AttachmentNg/c67477ae-0d96-4936-91ba-cd836bfa321e
Share
Share
Copy Link
AMD unveils its next-generation AI accelerator, the Instinct MI325X, along with new networking solutions, aiming to compete with Nvidia in the rapidly growing AI infrastructure market.
Advanced Micro Devices (AMD) has introduced its latest artificial intelligence (AI) chip, the Instinct MI325X accelerator, in a bold move to challenge Nvidia's dominance in the AI hardware market. The announcement came during AMD's "Advancing AI" event in San Francisco, where the company also revealed new networking solutions and updated its AI market projections 1.
The Instinct MI325X, built on AMD's CDNA 3 architecture, boasts impressive specifications:
AMD claims the MI325X delivers double the performance of its predecessor, the MI300, for AI inference and training workloads. Production is set to begin in Q4 2024, with widespread availability through major server manufacturers in early 2025 2.
Complementing the new accelerator, AMD introduced:
These networking solutions aim to optimize data transfer and enhance overall AI system performance, addressing critical bottlenecks in AI infrastructure 1.
AMD's aggressive push into the AI chip market has significant implications:
However, analysts remain cautious about AMD's ability to significantly disrupt Nvidia's market position in the short term, citing Nvidia's established ecosystem and software advantages 5.
AMD showcased collaborations with major tech players, including Google Cloud, Microsoft, Meta, and Oracle Cloud Infrastructure. These partnerships are crucial for AMD's strategy to gain traction in the AI infrastructure market 5.
As the AI hardware race intensifies, businesses across various sectors may benefit from increased competition, potentially leading to more accessible and cost-effective AI solutions in the long term. However, the immediate impact on pricing is expected to be limited due to the current high demand for AI chips 5.
Reference
[1]
[2]
[4]
[5]
AMD announces its new MI325X AI accelerator chip, set to enter mass production in Q4 2024, aiming to compete with Nvidia's upcoming Blackwell architecture in the rapidly growing AI chip market.
25 Sources
AMD reports strong Q3 results driven by AI-related growth, particularly in data center GPUs. Despite trailing Nvidia, AMD is carving out a significant position in the AI chip market with its MI300 series.
16 Sources
AMD's AI GPU business, led by the Instinct MI300, has grown rapidly to match the company's entire CPU operations in revenue. CEO Lisa Su predicts significant market growth, positioning AMD as a strong competitor to Nvidia in the AI hardware sector.
4 Sources
AMD's CEO Lisa Su emphasizes the company's accelerated AI roadmap and the ongoing AI industry growth. She discusses AMD's strategic positioning and future plans in the rapidly evolving AI market.
2 Sources
AMD reports strong Q2 2024 earnings, driven by exceptional AI chip sales and data center growth. The company's Instinct MI300 accelerators gain traction in the AI market, challenging NVIDIA's dominance.
3 Sources
The Outpost is a comprehensive collection of curated artificial intelligence software tools that cater to the needs of small business owners, bloggers, artists, musicians, entrepreneurs, marketers, writers, and researchers.
© 2024 TheOutpost.AI All rights reserved