Comparing energy consumption and accuracy in text classification inference
By: Johannes Zschache, Tilman Hartwig
Potential Business Impact:
Makes AI use less power to do tasks.
The increasing deployment of large language models (LLMs) in natural language processing (NLP) tasks raises concerns about energy efficiency and sustainability. While prior research has largely focused on energy consumption during model training, the inference phase has received comparatively less attention. This study systematically evaluates the trade-offs between model accuracy and energy consumption in text classification inference across various model architectures and hardware configurations. Our empirical analysis shows that the best-performing model in terms of accuracy can also be energy-efficient, while larger LLMs tend to consume significantly more energy with lower classification accuracy. We observe substantial variability in inference energy consumption ($<$mWh to $>$kWh), influenced by model type, model size, and hardware specifications. Additionally, we find a strong correlation between inference energy consumption and model runtime, indicating that execution time can serve as a practical proxy for energy usage in settings where direct measurement is not feasible. These findings have implications for sustainable AI development, providing actionable insights for researchers, industry practitioners, and policymakers seeking to balance performance and resource efficiency in NLP applications.
Similar Papers
Energy Considerations of Large Language Model Inference and Efficiency Optimizations
Computation and Language
Cuts AI's energy use by 73%.
Quantifying the Energy Consumption and Carbon Emissions of LLM Inference via Simulations
Distributed, Parallel, and Cluster Computing
Makes AI use less electricity and pollution.
TokenPowerBench: Benchmarking the Power Consumption of LLM Inference
Machine Learning (CS)
Measures AI's energy use to save power.