2 Sources
[1]
New chip uses AI to shrink large language models' energy footprint by 50%
"We have designed and fabricated a new chip that consumes half the energy compared to traditional designs," said doctoral student Ramin Javadi, who along with Tejasvi Anand, associate professor of electrical engineering, presented the technology at the recent IEEE Custom Integrated Circuits Conference in Boston. "The problem is that the energy required to transmit a single bit is not being reduced at the same rate as the data rate demand is increasing," said Anand, who directs the Mixed Signal Circuits and Systems Lab at OSU. "That's what is causing data centers to use so much power." The new chip itself is based on AI principles that reduce electricity use for signal processing, Javadi said. "Large language models need to send and receive tremendous amounts of data over wireline, copper-based communication links in data centers, and that requires significant energy," he said. "One solution is to develop more efficient wireline communication chips." When data is sent at high speeds, Javadi explains, it gets corrupted at the receiver and has to be cleaned up. Most conventional wireline communication systems use an equalizer to perform this task, and equalizers are comparatively power hungry. "We are using those AI principles on-chip to recover the data in a smarter and more efficient way by training the on-chip classifier to recognize and correct the errors," Javadi said. The Defense Advanced Research Projects Agency, the Semiconductor Research Corporation and the Center for Ubiquitous Connectivity supported the project, which earned Javadi the Best Student Paper Award at the conference. Javadi and Anand are working on the next iteration of the chip, which they expect to bring further gains in energy efficiency.
[2]
New chip uses AI to shrink large language models' energy footprint by 50%
Oregon State University College of Engineering researchers have developed a more efficient chip as an antidote to the vast amounts of electricity consumed by large-language-model artificial intelligence applications like Gemini and GPT-4. "We have designed and fabricated a new chip that consumes half the energy compared to traditional designs," said doctoral student Ramin Javadi, who, along with Tejasvi Anand, associate professor of electrical engineering, presented the technology at the IEEE Custom Integrated Circuits Conference in Boston. "The problem is that the energy required to transmit a single bit is not being reduced at the same rate as the data rate demand is increasing," said Anand, who directs the Mixed Signal Circuits and Systems Lab at OSU. "That's what is causing data centers to use so much power." The new chip itself is based on AI principles that reduce electricity use for signal processing, Javadi said. "Large language models need to send and receive tremendous amounts of data over wireline, copper-based communication links in data centers, and that requires significant energy," he said. "One solution is to develop more efficient wireline communication chips." When data is sent at high speeds, Javadi explains, it gets corrupted at the receiver and has to be cleaned up. Most conventional wireline communication systems use an equalizer to perform this task, and equalizers are comparatively power-hungry. "We are using those AI principles on-chip to recover the data in a smarter and more efficient way by training the on-chip classifier to recognize and correct the errors," Javadi said. Javadi and Anand are working on the next iteration of the chip, which they expect to bring further gains in energy efficiency.
Share
Copy Link
Oregon State University researchers have developed a chip that uses AI principles to reduce energy consumption by 50% in large language model applications, potentially revolutionizing data center efficiency.
Researchers at Oregon State University's College of Engineering have made a significant advancement in addressing the energy consumption challenges posed by large language models (LLMs) like Gemini and GPT-4. Doctoral student Ramin Javadi and Associate Professor Tejasvi Anand have developed a new chip that reduces energy consumption by 50% compared to traditional designs 12.
The rapid growth of data-intensive AI applications has led to a surge in energy consumption in data centers. Professor Anand, who directs the Mixed Signal Circuits and Systems Lab at OSU, explains the core issue: "The energy required to transmit a single bit is not being reduced at the same rate as the data rate demand is increasing. That's what is causing data centers to use so much power" 1.
The innovative chip leverages AI principles to optimize signal processing and data recovery. Javadi elaborates on the technology:
"Large language models need to send and receive tremendous amounts of data over wireline, copper-based communication links in data centers, and that requires significant energy. One solution is to develop more efficient wireline communication chips" 2.
The chip's efficiency stems from its novel approach to data correction:
The groundbreaking research has garnered attention and support from key organizations:
The OSU team is not resting on their laurels. Javadi and Anand are already working on the next iteration of the chip, anticipating even greater gains in energy efficiency 12. This ongoing research could have far-reaching implications for the future of AI infrastructure and data center operations.
If successfully implemented at scale, this technology could significantly reduce the carbon footprint of AI operations, making large language models more sustainable and cost-effective to run. It may also pave the way for more widespread adoption of AI technologies in various sectors by addressing one of the key challenges in AI deployment – energy consumption.
Summarized by
Navi
NASA and IBM have developed Surya, an open-source AI model that can predict solar flares and space weather, potentially improving the protection of Earth's critical infrastructure from solar storms.
5 Sources
Technology
7 hrs ago
5 Sources
Technology
7 hrs ago
Meta introduces an AI-driven voice translation feature for Facebook and Instagram creators, enabling automatic dubbing of content from English to Spanish and vice versa, with plans for future language expansions.
8 Sources
Technology
23 hrs ago
8 Sources
Technology
23 hrs ago
OpenAI CEO Sam Altman reveals plans for GPT-6, focusing on memory capabilities to create more personalized and adaptive AI interactions. The upcoming model aims to remember user preferences and conversations, potentially transforming the relationship between humans and AI.
2 Sources
Technology
23 hrs ago
2 Sources
Technology
23 hrs ago
Chinese AI companies DeepSeek and Baidu are making waves in the global AI landscape with their open-source models, challenging the dominance of Western tech giants and potentially reshaping the AI industry.
2 Sources
Technology
7 hrs ago
2 Sources
Technology
7 hrs ago
A comprehensive look at the emerging phenomenon of 'AI psychosis', its impact on mental health, and the growing concerns among experts and tech leaders about the psychological risks associated with AI chatbots.
3 Sources
Technology
7 hrs ago
3 Sources
Technology
7 hrs ago