ASE 2024
Sun 27 October - Fri 1 November 2024 Sacramento, California, United States
Wed 30 Oct 2024 10:30 - 10:45 at Gardenia - Code generation 2 Chair(s): Yangruibo Ding

Retrieval-augmented code generation utilizes Large Language Models as the generator and significantly expands their code generation capabilities by providing relevant code, documentation, and more via the retriever. The current approach suffers from two primary limitations: 1) \textbf{information redundancy.} The indiscriminate inclusion of redundant information can result in resource wastage and may misguide generators, affecting their effectiveness and efficiency. 2) \textbf{preference gap.} Due to different optimization objectives, the retriever strives to procure code with higher ground truth similarity, yet this effort does not substantially benefit the generator. The retriever and the generator may prefer different golden code, and this gap in preference results in a suboptimal design. Additionally, differences in parameterization knowledge acquired during pre-training result in varying preferences among different generators.

To address these limitations, in this paper, we propose \textbf{RRG} (\underline{\textbf{R}}etrieve, \underline{\textbf{R}}efactor, \underline{\textbf{G}}enerate), a novel framework for effective and efficient code generation. This framework introduces a code refactorer module between the retriever and the generator to bridge them. The refactoring process transforms the raw retrieved code into a more concise, efficient, and model-friendly version. It eliminates redundant information and noise, reducing the input length. Consequently, the generator receives higher-quality context, enabling it to produce more accurate results with lower inference costs. We conducted comprehensive experiments on multiple datasets. In the experiments, we confirmed the existence of a preference gap between the retriever and the generator, and RRG effectively bridges this gap. Specifically, RRG achieved significant performance improvements, with increases of up to 28% on EM, 13% on BLEU, and 6.8% on CodeBLEU.

Wed 30 Oct

Displayed time zone: Pacific Time (US & Canada) change

10:30 - 12:00
Code generation 2Research Papers / Tool Demonstrations at Gardenia
Chair(s): Yangruibo Ding Columbia University
10:30
15m
Talk
Preference-Guided Refactored Tuning for Retrieval Augmented Code Generation
Research Papers
Xinyu Gao , Yun Xiong Fudan University, Deze Wang National University of Defense Technology, Zhenhan Guan Fudan University, Zejian Shi Fudan University, Haofen Wang Tongji University, Shanshan Li National University of Defense Technology
Pre-print
10:45
15m
Talk
Sifting through the Chaff: On Utilizing Execution Feedback for Ranking the Generated Code Candidates
Research Papers
Zhihong Sun Shandong Normal University, Yao Wan Huazhong University of Science and Technology, Jia Li , Hongyu Zhang Chongqing University, Zhi Jin Peking University, Ge Li Peking University, Chen Lyu Shandong Normal University
11:00
15m
Talk
Promise and Peril of Collaborative Code Generation Models: Balancing Effectiveness and Memorization
Research Papers
Zhi Chen Singapore Management University, Lingxiao Jiang Singapore Management University
Pre-print
11:15
15m
Talk
JavaBench: A Benchmark of Object-Oriented Code Generation for Evaluating Large Language Models
Research Papers
Jialun Cao Hong Kong University of Science and Technology, Zhiyong Chen Nanjing University, Jiarong Wu The Hong Kong University of Science and Technology, Shing-Chi Cheung Hong Kong University of Science and Technology, Chang Xu Nanjing University
11:30
15m
Talk
PACGBI: A Pipeline for Automated Code Generation from Backlog Items
Tool Demonstrations
Mahja Sarschar Hochschule für Technik und Wirtschaft Berlin, Gefei Zhang HTW Berlin, Annika Nowak Capgemini
11:45
15m
Talk
Contextualized Data-Wrangling Code Generation in Computational Notebooks
Research Papers
Junjie Huang The Chinese University of Hong Kong, Daya Guo Sun-yat Sen University, Chenglong Wang Microsoft Research, Jiazhen Gu Chinese University of Hong Kong, Shuai Lu Microsoft Research, Jeevana Priya Inala Microsoft Research, Cong Yan Microsoft Research, Jianfeng Gao Microsoft Research, Nan Duan Microsoft Research, Michael Lyu The Chinese University of Hong Kong