2 Sources
2 Sources
[1]
New chip uses AI to shrink large language models' energy footprint by 50%
"We have designed and fabricated a new chip that consumes half the energy compared to traditional designs," said doctoral student Ramin Javadi, who along with Tejasvi Anand, associate professor of electrical engineering, presented the technology at the recent IEEE Custom Integrated Circuits Conference in Boston. "The problem is that the energy required to transmit a single bit is not being reduced at the same rate as the data rate demand is increasing," said Anand, who directs the Mixed Signal Circuits and Systems Lab at OSU. "That's what is causing data centers to use so much power." The new chip itself is based on AI principles that reduce electricity use for signal processing, Javadi said. "Large language models need to send and receive tremendous amounts of data over wireline, copper-based communication links in data centers, and that requires significant energy," he said. "One solution is to develop more efficient wireline communication chips." When data is sent at high speeds, Javadi explains, it gets corrupted at the receiver and has to be cleaned up. Most conventional wireline communication systems use an equalizer to perform this task, and equalizers are comparatively power hungry. "We are using those AI principles on-chip to recover the data in a smarter and more efficient way by training the on-chip classifier to recognize and correct the errors," Javadi said. The Defense Advanced Research Projects Agency, the Semiconductor Research Corporation and the Center for Ubiquitous Connectivity supported the project, which earned Javadi the Best Student Paper Award at the conference. Javadi and Anand are working on the next iteration of the chip, which they expect to bring further gains in energy efficiency.
[2]
New chip uses AI to shrink large language models' energy footprint by 50%
Oregon State University College of Engineering researchers have developed a more efficient chip as an antidote to the vast amounts of electricity consumed by large-language-model artificial intelligence applications like Gemini and GPT-4. "We have designed and fabricated a new chip that consumes half the energy compared to traditional designs," said doctoral student Ramin Javadi, who, along with Tejasvi Anand, associate professor of electrical engineering, presented the technology at the IEEE Custom Integrated Circuits Conference in Boston. "The problem is that the energy required to transmit a single bit is not being reduced at the same rate as the data rate demand is increasing," said Anand, who directs the Mixed Signal Circuits and Systems Lab at OSU. "That's what is causing data centers to use so much power." The new chip itself is based on AI principles that reduce electricity use for signal processing, Javadi said. "Large language models need to send and receive tremendous amounts of data over wireline, copper-based communication links in data centers, and that requires significant energy," he said. "One solution is to develop more efficient wireline communication chips." When data is sent at high speeds, Javadi explains, it gets corrupted at the receiver and has to be cleaned up. Most conventional wireline communication systems use an equalizer to perform this task, and equalizers are comparatively power-hungry. "We are using those AI principles on-chip to recover the data in a smarter and more efficient way by training the on-chip classifier to recognize and correct the errors," Javadi said. Javadi and Anand are working on the next iteration of the chip, which they expect to bring further gains in energy efficiency.
Share
Share
Copy Link
Oregon State University researchers have developed a chip that uses AI principles to reduce energy consumption by 50% in large language model applications, potentially revolutionizing data center efficiency.
Researchers at Oregon State University's College of Engineering have made a significant advancement in addressing the energy consumption challenges posed by large language models (LLMs) like Gemini and GPT-4. Doctoral student Ramin Javadi and Associate Professor Tejasvi Anand have developed a new chip that reduces energy consumption by 50% compared to traditional designs
1
2
.The rapid growth of data-intensive AI applications has led to a surge in energy consumption in data centers. Professor Anand, who directs the Mixed Signal Circuits and Systems Lab at OSU, explains the core issue: "The energy required to transmit a single bit is not being reduced at the same rate as the data rate demand is increasing. That's what is causing data centers to use so much power"
1
.The innovative chip leverages AI principles to optimize signal processing and data recovery. Javadi elaborates on the technology:
"Large language models need to send and receive tremendous amounts of data over wireline, copper-based communication links in data centers, and that requires significant energy. One solution is to develop more efficient wireline communication chips"
2
.The chip's efficiency stems from its novel approach to data correction:
1
2
.The groundbreaking research has garnered attention and support from key organizations:
1
Related Stories
The OSU team is not resting on their laurels. Javadi and Anand are already working on the next iteration of the chip, anticipating even greater gains in energy efficiency
1
2
. This ongoing research could have far-reaching implications for the future of AI infrastructure and data center operations.If successfully implemented at scale, this technology could significantly reduce the carbon footprint of AI operations, making large language models more sustainable and cost-effective to run. It may also pave the way for more widespread adoption of AI technologies in various sectors by addressing one of the key challenges in AI deployment – energy consumption.
Summarized by
Navi
1
Business and Economy
2
Business and Economy
3
Technology