Curated by THEOUTPOST
On Wed, 25 Sept, 12:07 AM UTC
7 Sources
[1]
Intel debuts super-fast AI chips - looking to match AMD in the race for AI chip supremacy
Intel has been having a challenging time in recent years, particularly following the launch of the iPhone with Apple silicon, and the rise of an AI and data center chip industry largely dominated by Nvidia and AMD. The company isn't taking these developments lying down, though, and has announced the release of its Xeon 6 chips and new Gaudi 3 AI accelerators. The Intel Xeon 6900 P-core series, as the chips are known, offers up to 128 cores for extremely intense AI workloads, doubling the performance of its predecessors, with higher core counts, more memory bandwidth, and embedded AI acceleration. On the Gaudi side, the new Gaudi 3 AI Accelerator chips, aimed at generative AI, come with 64 Tensor CPU cores, eight matrix multiplication engines, and 128GB of HBM2e memory, offering up to 20% more throughput and twice the price and performance over Nvidia's H100 chips for LLaMa 2 70B inference. The specs Intel is touting for the Gaudi 3 and Xeon 6 chips are really impressive, and would signal the company has managed to innovate at least a little out of the hole it finds itself in. Beyond designing and making the chips, Intel has also partnered with Dell and Supermicro to work on "co-engineered systems" that are made specifically for the specific AI needs of those companies. For its part, Dell is co-engineering RAG-based solutions leveraging Gaudi 3 and Xeon 6. AMD has been the biggest beneficiary of Intel's largesse in the AI chip space, designing powerful chips that are then manufactured by TSMC. It's a good business model and one that Intel has largely struggled to replicate or disrupt. The company plans to release its fifth-gen EPYC Turin 3nm data center chips in the second half of 2024 sometimes, based on its Zen 5 architecture and featuring up to 192 cores and 384 threads, so more than a match for the Xeon and Gaudi chips, at least on paper. In the closely-watched Top500 list, the Intel-powered Aurora supercomputer, which is not yet fully operational, came second behind the AMD-powered Frontier, but did take first in an AI benchmark.
[2]
Intel launches Xeon 6 and Gaudi 3 AI chips to boost AI and HPC performance
GamesBeat Next is connecting the next generation of video game leaders. And you can join us, coming up October 28th and 29th in San Francisco! Take advantage of our buy one, get one free pass offer. Sale ends this Friday, August 16th. Join us by registering here. Intel is launching new Xeon 6 processors with performance cores as well as Gaudi 3 AI accelerators to stay competitive in the AI wars. The new Xeon 6 processors have performance cores (P-cores) that can double AI vision performance and the Gaudi 3 AI accelerators have 20% more throughput. As AI continues to revolutionize industries, enterprises are increasingly in need of infrastructure that is both cost-effective and available for rapid development and deployment. To meet this demand head-on, Intel today launched Xeon 6 with Performance-cores (P-cores) and Gaudi 3 AI accelerators, bolstering the company's commitment to deliver powerful AI systems with optimal performance per watt and lower total cost of ownership (TCO). "Demand for AI is leading to a massive transformation in the data center, and the industry is asking for choice in hardware, software, and developer tools," said Justin Hotard, Intel executive vice president and general manager of the data center and AI group at Intel, in a statement. "With our launch of Xeon 6 with P-cores and Gaudi 3 AI accelerators, Intel is enabling an open ecosystem that allows our customers to implement all of their workloads with greater performance, efficiency, and security." Introducing Intel Xeon 6 with P-cores and Gaudi 3 AI accelerators Intel's latest advancements in AI infrastructure include two major updates to its data center portfolio. These include Intel Xeon6 with P-cores. They're designed to handle compute-intensive workloads with exceptional efficiency, Xeon 6 delivers twice the performance of its predecessor. It features increased core count, double the memory bandwidth and AI acceleration capabilities embedded in every core. This processor is engineered to meet the performance demands of AI from edge to data center and cloud environments. The Intel Gaudi 3 AI accelerator is specifically optimized for large-scale generative AI, Gaudi 3 boasts 64 Tensor processor cores (TPCs) and eight matrix multiplication engines (MMEs) to accelerate deep neural network computations. It includes 128 gigabytes (GB) of HBMe2 memory for training and inference, and 24 200 Gigabit (Gb) Ethernet ports for scalable networking. Gaudi 3 also offers seamless compatibility with the PyTorch framework and advanced Hugging Face transformer and diffuser models. Intel recently announced a collaboration with IBM to deploy Intel Gaudi 3 AI accelerators as a service on IBM Cloud. Through this collaboration, Intel and IBM aim to lower the total cost of ownership to leverage and scale AI, while enhancing performance. Enhancing AI systems with TCO benefits Deploying AI at scale involves considerations such as flexible deployment options, competitive price-performance ratios and accessible AI technologies. Intel's robust x86 infrastructure and extensive open ecosystem position it to support enterprises in building high-value AI systems with an optimal TCO and performance per watt. Notably, 73% of GPU-accelerated servers use Intel Xeon as the host CPU. Intel has partnered with with leading original equipment manufacturers (OEMs) including Dell Technologies, Hewlett Packard Enterprise, and Supermicro to develop co-engineered systems tailored to specific customer needs for effective AI deployments. Dell Technologies is currently co-engineering RAG-based solutions leveraging Gaudi 3 and Xeon 6. Transitioning generative AI (Gen AI) solutions from prototypes to production-ready systems presents challenges in real-time monitoring, error handling, logging, security and scalability. Intel addresses these challenges through co-engineering efforts with OEMs and partners to deliver production-ready retrieval-augmented generation (RAG) solutions. These solutions, built on the Open Platform Enterprise AI (OPEA) platform, integrate OPEA-based microservices into a scalable RAG system, optimized for Xeon and Gaudi AI systems, designed to allow customers to easily integrate applications from Kubernetes and Red Hat OpenShift. Expanding access to enterprise AI applications Intel's Tiber portfolio offers business solutions to tackle challenges such as access, cost, complexity, security, efficiency and scalability across AI, cloud and edge environments. The Intel® Tiber™ Developer Cloud now provides preview systems of Intel Xeon 6 for tech evaluation and testing. Additionally, select customers will gain early access to Intel Gaudi 3 for validating AI model deployments, with Gaudi 3 clusters to begin rolling out next quarter for large-scale production deployments. New service offerings include SeekrFlow, an end-to-end AI platform from Seekr for developing trusted AI applications. The latest updates feature Intel Gaudi software's newest release and Jupyter notebooks loaded with PyTorch 2.4 and Intel oneAPI and AI tools 2024.2, which include new AI acceleration capabilities and support for Xeon 6 processors.
[3]
Intel Unveils Next-Generation AI Solutions with the Launch of Xeon 6 and Gaudi 3
Intel launched Xeon 6 with Performance-cores (P-cores) and Gaudi 3 AI accelerators, bolstering the company's commitment to deliver powerful AI systems with optimal performance per watt and lower total cost of ownership (TCO). Introducing Intel Xeon 6 with P-cores and Gaudi 3 AI accelerators Intel's latest advancements in AI infrastructure include two major updates to its data center portfolio: Intel® Xeon® 6 with P-cores: Designed to handle compute-intensive workloads with exceptional efficiency, Xeon 6 delivers twice the performance of its predecessor2. It features increased core count, double the memory bandwidth and AI acceleration capabilities embedded in every core. This processor is engineered to meet the performance demands of AI from edge to data center and cloud environments. Intel® Gaudi® 3 AI Accelerator: Specifically optimized for large-scale generative AI, Gaudi 3 boasts 64 Tensor processor cores (TPCs) and eight matrix multiplication engines (MMEs) to accelerate deep neural network computations. It includes 128 gigabytes (GB) of HBM2e memory for training and inference, and 24 200 Gigabit (Gb) Ethernet ports for scalable networking. Gaudi 3 also offers seamless compatibility with the PyTorch framework and advanced Hugging Face transformer and diffuser models. Intel recently announced a collaboration with IBM to deploy Intel Gaudi 3 AI accelerators as a service on IBM Cloud. Through this collaboration, Intel and IBM aim to lower the total cost of ownership to leverage and scale AI, while enhancing performance. Enhancing AI Systems with TCO Benefits Deploying AI at scale involves considerations such as flexible deployment options, competitive price-performance ratios and accessible AI technologies. Intel's robust x86 infrastructure and extensive open ecosystem position it to support enterprises in building high-value AI systems with an optimal TCO and performance per watt. Notably, 73% of GPU-accelerated servers use Intel Xeon as the host CPU. Intel partners with leading OEMs including Dell Technologies and Supermicro to develop co-engineered systems tailored to specific customer needs for effective AI deployments. Dell Technologies is currently co-engineering RAG-based solutions leveraging Gaudi 3 and Xeon 6.
[4]
Intel Launches Xeon 6 CPUs and Gaudi 3 AI Accelerators to Enhance Data Center AI Performance
Intel has unveiled its latest advancements in data center technology with the introduction of the Xeon 6 processors and Gaudi 3 AI accelerators, designed to meet the increasing demands of artificial intelligence (AI) workloads. The Xeon 6 processors feature Performance-cores (P-cores) that significantly enhance computational efficiency, delivering twice the performance of the previous generation. These processors boast an increased core count and doubled memory bandwidth, alongside integrated AI acceleration within each core. This architecture is optimized to handle compute-intensive tasks across various environments, including edge devices, data centers, and cloud infrastructures, thereby providing a robust foundation for diverse AI applications. Complementing the Xeon 6 processors, Intel's Gaudi 3 AI accelerators are specifically engineered for large-scale generative AI tasks. The Gaudi 3 is equipped with 64 Tensor Processor Cores (TPCs) and eight Matrix Multiplication Engines (MMEs), which facilitate accelerated deep neural network computations. It includes 128 gigabytes of HBM2e memory, ensuring efficient data handling for both training and inference processes. Additionally, the accelerator offers 24 200 Gigabit Ethernet ports to support scalable networking needs. The Gaudi 3 is fully compatible with the PyTorch framework and supports advanced models from Hugging Face, enabling seamless integration with existing AI development tools. In collaboration with IBM, Intel is deploying Gaudi 3 accelerators as a service on IBM Cloud, aiming to reduce the total cost of ownership (TCO) and enhance the scalability of AI deployments for enterprises. Intel's introduction of the Xeon 6 CPUs and Gaudi 3 AI accelerators underscores its commitment to optimizing TCO and performance per watt through its established x86 infrastructure and extensive open ecosystem. Currently, 73% of GPU-accelerated servers utilize Intel Xeon as the host CPU, highlighting its widespread adoption in the industry. Partnerships with original equipment manufacturers (OEMs) such as Dell Technologies and Supermicro enable the development of co-engineered systems tailored to specific customer requirements, facilitating effective AI implementations. Intel also addresses the challenges of transitioning generative AI solutions from prototypes to production-ready systems through collaborative engineering efforts and the Open Platform Enterprise AI (OPEA) framework. This framework integrates scalable retrieval-augmented generation (RAG) systems optimized for Xeon and Gaudi accelerators, allowing customers to easily integrate applications from platforms like Kubernetes and Red Hat OpenShift AI. Additionally, Intel's Tiber portfolio provides comprehensive solutions that tackle issues related to access, cost, complexity, security, efficiency, and scalability across AI, cloud, and edge environments. The Intel Tiber Developer Cloud offers preview systems of Xeon 6 for technical evaluation and testing, while select customers receive early access to Gaudi 3 accelerators to validate AI model deployments ahead of large-scale production rollouts. New service offerings, such as the SeekrFlow AI platform, incorporate the latest Intel Gaudi software and tools, including support for Xeon 6 processors, to facilitate the development of reliable and efficient AI applications.
[5]
Intel launches Xeon 6 for AI and high-performance computing, Gaudi 3 AI accelerators - Times of India
Intel has launched its latest offerings in AI infrastructure: the Xeon 6 processor with Performance-cores (P-cores) and the Gaudi 3 AI accelerator. The company aims to provide powerful AI systems with improved performance per watt and lower total cost of ownership (TCO). As per Intel, the Xeon 6 processor is built for compute-intensive tasks, offering twice the performance of its predecessor.It has more cores, double the memory bandwidth, and built-in AI acceleration in every core, the company said. This processor is engineered to meet the performance demands of AI from edge to data centre and cloud environments. The Gaudi 3 AI accelerator, on the other hand, is optimised for large-scale generative AI. It has 64 Tensor processor cores and 128GB of memory. It is compatible with PyTorch and Hugging Face models, and Intel has partnered with IBM to offer it as a service on IBM Cloud. Intel has also announced that it is collaborating with OEMs like Dell Technologies to create tailored solutions for specific customer needs. Intel's Lunar Lake processors The launch comes about 20 days after Intel launched the Lunar Lake processors. The laptops with the Intel Core Ultra 200V chips have started to hit shelves. Intel says that the new processors are aimed at meeting the consumers' demand for powerful performance, long-lasting battery life, wide application compatibility and enhanced security, along with Al support. Intel claims that the new processors can deliver up to 20.1 hours battery life and are said to offer up to 120 total platform TOPS (tera operations per second) across central processing unit (CPU), graphic processing unit (GPU) and neural processing unit (NPU) for AI performance. The TOI Tech Desk is a dedicated team of journalists committed to delivering the latest and most relevant news from the world of technology to readers of The Times of India. TOI Tech Desk's news coverage spans a wide spectrum across gadget launches, gadget reviews, trends, in-depth analysis, exclusive reports and breaking stories that impact technology and the digital universe. Be it how-tos or the latest happenings in AI, cybersecurity, personal gadgets, platforms like WhatsApp, Instagram, Facebook and more; TOI Tech Desk brings the news with accuracy and authenticity.
[6]
Intel : Tackles Enterprise Needs with AI Systems
SANTA CLARA, Calif., September 24, 2024 - As AI continues to revolutionize industries, enterprises are increasingly in need of infrastructure that is both cost-effective and available for rapid development and deployment. To meet this demand head-on, Intel today launched Xeon 6 with Performance-cores (P-cores) and Gaudi 3 AI accelerators, bolstering the company's commitment to deliver powerful AI systems with optimal performance per watt and lower total cost of ownership (TCO). "Demand for AI is leading to a massive transformation in the data center, and the industry is asking for choice in hardware, software and developer tools," said Justin Hotard, Intel executive vice president and general manager of the Data Center and Artificial Intelligence Group. "With our launch of Xeon 6 with P-cores and Gaudi 3 AI accelerators, Intel is enabling an open ecosystem that allows our customers to implement all of their workloads with greater performance, efficiency and security." More:Intel Xeon 6 with P-cores and Gaudi 3 AI Accelerators (Press Kit) | Leadership performance with Intel Xeon 6900 P-core series (Fact Sheet) | A New Era of High-Performance Enterprise AI Systems and Solutions (Quote Sheet) Introducing Intel Xeon 6 with P-cores and Gaudi 3 AI accelerators Intel's latest advancements in AI infrastructure include two major updates to its data center portfolio: Enhancing AI Systems with TCO Benefits Deploying AI at scale involves considerations such as flexible deployment options, competitive price-performance ratios and accessible AI technologies. Intel's robust x86 infrastructure and extensive open ecosystem position it to support enterprises in building high-value AI systems with an optimal TCO and performance per watt. Notably, 73% of GPU-accelerated servers use Intel Xeon as the host CPU. Intel partners with leading OEMs including Dell Technologies and Supermicro to develop co-engineered systems tailored to specific customer needs for effective AI deployments. Dell Technologies is currently co-engineering RAG-based solutions leveraging Gaudi 3 and Xeon 6.
[7]
Intel launches Xeon 6 and Gaudi 3 AI accelerators to boost AI performance
Will AI replace programmers? Don't count on it, says Google's CEO Intel just unveiled its newest hardware aimed at enhancing AI infrastructure: the Xeon 6 with Performance-cores (P-cores) and the Gaudi 3 AI accelerator. These latest additions are designed to provide high-performance, cost-effective options for businesses looking to scale up their AI operations. Related 10 ways you can use AI to improve your productivity From help managing your schedule, to writing emails, AI can make you a lot more productive Xeon 6: More Power for AI The Xeon 6 is a big upgrade, offering twice the performance of its predecessor. Intel has packed in more cores, double the memory bandwidth, and built-in AI acceleration. The idea is to make Xeon 6 versatile enough to handle various workloads, from edge computing to large data centers. Whether you're crunching numbers for machine learning or powering up AI models, Xeon 6 is built to keep things running smoothly across a range of environments. Gaudi 3: Supercharging AI Tasks For even more muscle in AI tasks, Intel's Gaudi 3 AI accelerator comes into play. Designed for massive, compute-heavy jobs like generative AI, it features 64 Tensor cores and eight matrix multiplication engines. Translation: it's built to process neural network computations fast. With 128GB of HBM2e memory and a hefty 24 200-Gigabit Ethernet ports, it can handle large-scale tasks without breaking a sweat. Gaudi 3 is also optimized for frameworks like PyTorch and advanced AI models from Hugging Face. For companies running deep learning models, this accelerator could make things easier and faster. Intel has already partnered with IBM Cloud to offer Gaudi 3 as a service, making it more accessible for businesses that need scalable AI solutions. Focus on Efficiency and Cost Close Along with raw performance, Intel is pitching the cost-efficiency of these new products. Both the Xeon 6 and Gaudi 3 are built to offer a solid balance of performance per watt, which means they could lower operational costs for companies running AI workloads at scale. This fits into Intel's broader plan to provide more flexible AI infrastructure without driving up the total cost of ownership. Intel's partnership with Dell Technologies and Supermicro further highlights their strategy to integrate these products into customizable AI solutions. Whether you're running a smaller operation or a larger enterprise, Intel's aiming to provide AI tools that fit your specific needs. "Demand for AI is leading to a massive transformation in the data center, and the industry is asking for choice in hardware, software and developer tools," said Justin Hotard, Intel executive vice president and general manager of the Data Center and Artificial Intelligence Group. "With our launch of Xeon 6 with P-cores and Gaudi 3 AI accelerators, Intel is enabling an open ecosystem that allows our customers to implement all of their workloads with greater performance, efficiency and security." Related 3 reasons why Arm's AI work is the most interesting happening for consumers right now Arm has been putting a lot of work into AI that you actually care about.
Share
Share
Copy Link
Intel launches new Xeon 6 CPUs and Gaudi 3 AI accelerators to boost AI and high-performance computing capabilities in data centers, aiming to compete with AMD and NVIDIA in the AI chip market.
Intel, a leading player in the semiconductor industry, has made a significant leap in the artificial intelligence (AI) chip market with the launch of its next-generation AI solutions. The company has unveiled two new products: the Xeon 6 CPUs and the Gaudi 3 AI accelerators, both designed to enhance AI and high-performance computing (HPC) capabilities in data centers 1.
The new Xeon 6 CPUs are built on Intel's advanced manufacturing process and are specifically tailored for AI and HPC workloads. These processors offer improved performance and energy efficiency compared to their predecessors. Intel claims that the Xeon 6 CPUs provide up to 42% better performance for AI inference workloads and up to 36% higher performance for technical computing applications 2.
Alongside the Xeon 6 CPUs, Intel has introduced the Gaudi 3 AI accelerators. These specialized chips are designed to accelerate AI training and inference tasks. The Gaudi 3 accelerators boast significant performance improvements, with Intel claiming up to 4x faster AI training and up to 3x higher inference performance compared to their predecessors 3.
Intel's latest offerings are aimed at strengthening its position in the highly competitive AI chip market, where it faces stiff competition from rivals such as AMD and NVIDIA. The company is leveraging its expertise in CPU design and manufacturing to create AI-optimized solutions that can compete with specialized AI chips from other manufacturers 4.
The launch of Xeon 6 CPUs and Gaudi 3 AI accelerators is expected to have a significant impact on various industries. These new chips are designed to support a wide range of AI applications, including natural language processing, computer vision, and large language models. Intel's solutions aim to provide the necessary computational power for developing and deploying advanced AI systems across different sectors 5.
In addition to performance improvements, Intel has emphasized the energy efficiency of its new AI solutions. The company claims that the Xeon 6 CPUs and Gaudi 3 accelerators offer better performance per watt compared to previous generations, addressing the growing concern of power consumption in data centers 2.
As the demand for AI computing continues to grow, Intel's latest offerings position the company to capture a larger share of the AI chip market. The success of these new products could potentially reshape the competitive landscape in the semiconductor industry, particularly in the rapidly evolving field of AI hardware 1.
Reference
[1]
[4]
Intel launches new Xeon 6 processors with performance cores, offering improved AI processing and networking solutions for data centers and edge computing.
10 Sources
10 Sources
Intel has announced its latest high-performance server processors, the Xeon 6900P series, featuring up to 128 cores and significant performance improvements. This launch marks a major step in Intel's efforts to compete in the data center market.
3 Sources
3 Sources
Intel has introduced its new Gaudi 3 AI accelerator, positioning it as a more affordable option compared to NVIDIA's H100 GPU. While it offers lower performance, the Gaudi 3 aims to provide a cost-effective solution for AI workloads.
3 Sources
3 Sources
Intel launches Tiber AI Cloud, powered by Gaudi 3 chips, partnering with Inflection AI to offer enterprise AI solutions, competing with major cloud providers and NVIDIA in the AI accelerator market.
4 Sources
4 Sources
AMD announces its new MI325X AI accelerator chip, set to enter mass production in Q4 2024, aiming to compete with Nvidia's upcoming Blackwell architecture in the rapidly growing AI chip market.
25 Sources
25 Sources
The Outpost is a comprehensive collection of curated artificial intelligence software tools that cater to the needs of small business owners, bloggers, artists, musicians, entrepreneurs, marketers, writers, and researchers.
© 2025 TheOutpost.AI All rights reserved