Efficient and Green Large Language Models for Software Engineering: Literature Review, Vision, and the Road Ahead
This program is tentative and subject to change.
Large Language Models (LLMs) have demonstrated remarkable capabilities across a wide spectrum of software engineering tasks, from code generation and summarization to program repair and vulnerability detection. This surge has given rise to the Large Language Models for Software Engineering (LLM4SE) area. However, limited attention has been paid to address two critical issues of LLM4SE techniques: (1) \emph{inefficiency}: they require prohibitive computational cost, time, and memory resources, making them inaccessible to many practitioners, and (2) \emph{unsustainability}: they consume vast amounts of energy, water, and carbon emissions, raising serious environmental concerns. As a result, the current trajectory of LLM4SE risks becoming both economically and ecologically unsustainable, calling for urgent attention to efficiency and greenness.
This paper aims to redirect the focus of the research community towards the efficiency and greenness of LLM4SE, while also sharing potential research directions to achieve this goal. It commences with a brief overview of the significance of LLM4SE and highlights the need for efficient and green LLM4SE solutions, followed by a classification of existing techniques along four complementary dimensions: data-centric, modelcentric, system-centric, and program-centric. Building on this analysis, the paper presents a vision for a future where efficient and green LLM4SE revolutionizes the LLM-based software engineering tool landscape, benefiting various stakeholders, including industry, individual practitioners, and society. The paper then delineates a roadmap for future research, outlining several promising research directions to realize the above vision by 2030, such as constructing new benchmarks, developing efficient training and fine-tuning methods, exploring alternative approaches like retrieval-augmented generation, and advancing model compression, inference acceleration, and program-level optimization. While not intended to be a definitive guide, the paper aims to inspire further progress, with the ultimate goal of establishing efficient and green LLM4SE as a central element in the future of software engineering.
This paper has been accepted for publication in ACM Transactions on Software Engineering and Methodology (TOSEM) on November 07, 2024, and is available at https://doi.org/10.1145/3708525. Rather than being exclusively a secondary study, this paper combines a literature review with a forward-looking vision and detailed research roadmap. Jieke Shi will present the paper at ICSE 2026, if invited.
This program is tentative and subject to change.
Thu 16 AprDisplayed time zone: Brasilia, Distrito Federal, Brazil change
11:00 - 12:30 | AI for Software Engineering 10Research Track / Journal-first Papers at Asia I Chair(s): Fabio Marcos De Abreu Santos Colorado State University, USA | ||
11:00 15mTalk | FlipFlop: A Static Analysis-based Energy Optimization Framework for GPU Kernels Research Track Saurabhsingh Rajput Dalhousie University, Alexander Brandt Dalhousie University, Vadim Elisseev IBM, Tushar Sharma Dalhousie University | ||
11:15 15mTalk | Portable Power Modeling with Transfer Learning on JVM-Based Applications Research Track | ||
11:30 15mTalk | End-to-End Model Generation with Large Language Models for Adaptive IoT Application Deployment Research Track ZHENYU WEN Zhejiang University of Technology, Jintao Feng Zhejiang University of Technology, Yao Nanjie Zhejiang University of Technology, Di Wu University of Central Florida, Cong Wang Zhejiang University, China, Mincheng Wu Zhejiang University of Technology, Jianbin Qin Shenzhen Institute of Computing Sciences, Shenzhen University, Shibo He Zhejiang University | ||
11:45 15mTalk | Efficient and Green Large Language Models for Software Engineering: Literature Review, Vision, and the Road Ahead Journal-first Papers | ||
12:00 15mTalk | An Empirical Study of Knowledge Distillation for Code Understanding Tasks Research Track Ruiqi Wang Harbin Institute of Technology, Shenzhen, Zezhou Yang , Cuiyun Gao Harbin Institute of Technology, Shenzhen, Xin Xia Zhejiang University, Qing Liao Harbin Institute of Technology Pre-print | ||
12:15 15mTalk | Generating Energy-Efficient Code via Large-Language Models - Where are we now? Research Track Radu Apsan Vrije Universiteit Amsterdam, The Netherlands, Vincenzo Stoico Vrije Universiteit Amsterdam, Michel Albonico Federal University of Technology, Paraná (UTFPR), Rudra Dhar IIIT Hyderabad, Karthik Vaidhyanathan IIIT Hyderabad, Ivano Malavolta Vrije Universiteit Amsterdam Pre-print Media Attached | ||