2 Sources
2 Sources
[1]
IBM partners with Nvidia rival Groq to accelerate AI deployment - SiliconANGLE
IBM partners with Nvidia rival Groq to accelerate AI deployment IBM Corp. and Groq Inc. today announced a strategic partnership aimed at speeding enterprise deployment of agentic artificial intelligence by combining IBM's watsonx Orchestrate with Groq's hardware-accelerated inference technology. The collaboration will give IBM clients access to Groq's language processing unit through IBM's platform with the aim of reducing the cost of low-latency AI at large scale. Groq is a hardware-centric AI infrastructure company specializing in inference acceleration through its proprietary LPU architecture. IBM said Groq's custom LPU delivers more than five times faster and more cost-efficient inference than traditional graphics processing units such as those made by Nvidia Corp., maintaining consistent performance even as workloads scale up. The companies will also extend support for virtual large language models and Red Hat's llm-d framework for large-scale, distributed inference on Groq's LPU architecture. IBM's Granite models will run on GroqCloud, an inference platform that provides developers with access to high-speed, low-cost processing for large language models. IBM said Groq provides a secure, compliant platform for AI deployment that meets the most stringent regulatory and security requirements. Watsonx Orchestrate is used to automate business processes using prebuilt or custom-designed AI agents. Integration with watsonx Orchestrate is aimed at helping IBM customers speed purpose-built agents to production faster and at lower cost. Groq has raised $1.8 billion in funding, including a $750 million round last month that valued the company at $6.9 billion. IBM said a combination of speed and orchestration is critical for regulated industries such as healthcare and finance. It cited one health insurance provider that uses Groq technology to handle thousands of patient and provider questions simultaneously and in real time. "Many large enterprise organizations have a range of options with AI inferencing when they're experimenting, but when they want to go into production, they must ensure complex workflows can be deployed successfully to ensure high-quality experiences," Rob Thomas, IBM senior vice president of software and chief commercial officer, said in a statement. IBM said access to Groq's capabilities is available immediately.
[2]
IBM partners with Groq in agentic AI
On Monday, IBM announced the signing of a strategic and technological partnership with the Groq AI platform to enable its customers to access inference workloads on GroqCloud from watsonx Orchestrate, its agentic AI tool. Under the agreement, IBM plans to integrate RedHat's large open-source virtual language model with Groq's language processing unit (LPU) architecture. IBM Granite language models will also be supported on GroqCloud for IBM customers. IBM explains that the project aims to facilitate the transition from the pilot phase of agentic AI to production in areas such as healthcare, finance, public administration, distribution, and manufacturing by reducing costs and accelerating deployment.
Share
Share
Copy Link
IBM and Groq announce a strategic partnership to enhance AI deployment for enterprises. The collaboration aims to speed up the implementation of agentic AI by integrating Groq's hardware-accelerated inference technology with IBM's watsonx Orchestrate platform.

IBM and Groq, a rising star in AI hardware acceleration, have announced a strategic partnership that promises to reshape the landscape of enterprise AI deployment. This collaboration aims to accelerate the implementation of agentic artificial intelligence by combining IBM's watsonx Orchestrate with Groq's cutting-edge hardware-accelerated inference technology
1
.At the heart of this partnership is the integration of Groq's Language Processing Unit (LPU) architecture with IBM's established AI platforms. IBM clients will gain access to Groq's LPU through the watsonx Orchestrate platform, potentially revolutionizing the speed and cost-effectiveness of AI deployment at scale
1
.Groq's proprietary LPU technology is reported to deliver more than five times faster and more cost-efficient inference compared to traditional graphics processing units (GPUs), such as those produced by industry leader Nvidia. A key advantage of Groq's architecture is its ability to maintain consistent performance even as workloads increase in scale
1
.The collaboration extends beyond hardware integration. IBM and Groq are working to support virtual large language models and Red Hat's llm-d framework for large-scale, distributed inference on Groq's LPU architecture. Additionally, IBM's Granite models will be made available on GroqCloud, an inference platform that offers developers high-speed, low-cost processing for large language models
1
.IBM emphasizes that this partnership is particularly beneficial for regulated industries such as healthcare and finance, where a combination of speed and orchestration is critical. The integration of Groq's technology with watsonx Orchestrate is expected to help IBM customers accelerate the deployment of purpose-built AI agents to production faster and at a lower cost
2
.Rob Thomas, IBM's Senior Vice President of Software and Chief Commercial Officer, highlighted the importance of this collaboration for large enterprise organizations looking to move from AI experimentation to production. He stressed the need for successfully deploying complex workflows to ensure high-quality experiences
1
.Related Stories
This partnership comes at a time when Groq is gaining significant traction in the AI hardware market. The company has raised an impressive $1.8 billion in funding, including a recent $750 million round that valued Groq at $6.9 billion
1
.IBM has announced that access to Groq's capabilities is immediately available to its clients, marking a swift implementation of this strategic partnership
1
. This rapid deployment underscores the urgency and potential impact of this collaboration on the enterprise AI landscape.Summarized by
Navi
[2]
25 Nov 2024•Technology

07 Oct 2025•Technology

06 May 2025•Technology

1
Technology

2
Business and Economy

3
Business and Economy
