Energy-Aware LLMs: A step towards sustainable AI for downstream applications
By: Nguyen Phuc Tran, Brigitte Jaumard, Oscar Delgado
Potential Business Impact:
Saves energy while making AI smarter.
Advanced Large Language Models (LLMs) have revolutionized various fields, including communication networks, sparking an innovation wave that has led to new applications and services, and significantly enhanced solution schemes. Despite all these impressive developments, most LLMs typically require huge computational resources, resulting in terribly high energy consumption. Thus, this research study proposes an end-to-end pipeline that investigates the trade-off between energy efficiency and model performance for an LLM during fault ticket analysis in communication networks. It further evaluates the pipeline performance using two real-world datasets for the tasks of root cause analysis and response feedback in a communication network. Our results show that an appropriate combination of quantization and pruning techniques is able to reduce energy consumption while significantly improving model performance.
Similar Papers
Optimizing Large Language Models: Metrics, Energy Efficiency, and Case Study Insights
Machine Learning (CS)
Cuts AI's energy use by almost half.
Large Language Models integration in Smart Grids
Computers and Society
Makes power grids smarter and more reliable.
From Large to Super-Tiny: End-to-End Optimization for Cost-Efficient LLMs
Computation and Language
Makes smart computer programs cheaper and faster.