International Journal of Computer Applications |
Foundation of Computer Science (FCS), NY, USA |
Volume 187 - Number 14 |
Year of Publication: 2025 |
Authors: Krishnam Raju Narsepalle |
![]() |
Krishnam Raju Narsepalle . Energy-Efficient Training and Inference in Large Language Models: Optimizing Computational and Energy Costs. International Journal of Computer Applications. 187, 14 ( Jun 2025), 1-13. DOI=10.5120/ijca2025925323
The larger the size of the Large Language Models (LLMs) is, the higher their computational and energy costs become, and thus, the environmental and economic impact increases. This paper examines several initiatives aimed at reducing the energy and computational costs associated with training and deploying Large Language Models (LLMs). Training sparse, adaptive inference, and hardware acceleration (based on GPUs and TPUs) are assessed. The modelling experiments using BERT and GPT indicate that sparse training reduces the computational workload by an additional 35%, while adaptive inference significantly reduces energy consumption during inference by 20%. Additionally, a 25% energy savings has been achieved by optimizing resource loading on the hardware. These findings suggest that energy-efficient Large Language Model (LLM) training and inference methods can significantly reduce the environmental impact of large-scale AI models, making them more sustainable for widespread use.