ASE 2024
Sun 27 October - Fri 1 November 2024 Sacramento, California, United States

This program is tentative and subject to change.

Wed 30 Oct 2024 11:00 - 11:15 at Gardenia - Code generation 2

In the rapidly evolving field of machine learning, training models with datasets from various locations and organizations presents significant challenges due to privacy and legal concerns. The exploration of effective collaborative training settings, which are capable of leveraging valuable knowledge from distributed and isolated datasets, is increasingly crucial. This study investigates key factors that impact the effectiveness of collaborative training methods in code next-token prediction, as well as the correctness and utility of the generated code, showing the promise of such methods. Additionally, we evaluate the memorization of different participant training data across various collaborative training settings, including centralized, federated, and incremental training, showing their potential risks in leaking data.

Our findings indicate that the size and diversity of code datasets are pivotal factors influencing the success of collaborative trained code models. We demonstrate that federated learning achieves competitive performance compared to centralized training while offering better data protection, as evidenced by lower memorization ratios in the generated code. However, federated learning can still produce verbatim code snippets from hidden training data, potentially violating data privacy or copyright. Our study further explores the patterns of effectiveness and memorization in incremental learning, emphasizing the importance of the sequence in which individual participant datasets are introduced. Also, we identify the memorization phenomenon of cross-organizational clones as a prevalent challenge in both centralized and federated learning scenarios. Our findings highlight the persistent risk of data leakage during inference, even when training data remains unseen. We conclude with strategic recommendations for practitioners and researchers to optimize the use of multisource datasets, thereby propelling the cross-organizational collaboration forward.

This program is tentative and subject to change.

Wed 30 Oct

Displayed time zone: Pacific Time (US & Canada) change

10:30 - 12:00
10:30
15m
Talk
Preference-Guided Refactored Tuning for Retrieval Augmented Code Generation
Research Papers
Xinyu Gao , Yun Xiong Fudan University, Deze Wang National University of Defense Technology, Zhenhan Guan Fudan University, Zejian Shi Fudan University, Haofen Wang Tongji University, Shanshan Li National University of Defense Technology
Pre-print
10:45
15m
Talk
Sifting through the Chaff: On Utilizing Execution Feedback for Ranking the Generated Code Candidates
Research Papers
Zhihong Sun Shandong Normal University, Yao Wan Huazhong University of Science and Technology, Jia Li , Hongyu Zhang Chongqing University, Zhi Jin Peking University, Ge Li Peking University, Chen Lyu Shandong Normal University
11:00
15m
Talk
Promise and Peril of Collaborative Code Generation Models: Balancing Effectiveness and Memorization
Research Papers
Zhi Chen Singapore Management University, Lingxiao Jiang Singapore Management University
Pre-print
11:15
15m
Talk
JavaBench: A Benchmark of Object-Oriented Code Generation for Evaluating Large Language Models
Research Papers
Jialun Cao Hong Kong University of Science and Technology, Zhiyong Chen Nanjing University, Jiarong Wu The Hong Kong University of Science and Technology, Shing-Chi Cheung Hong Kong University of Science and Technology, Chang Xu Nanjing University
11:30
15m
Talk
PACGBI: A Pipeline for Automated Code Generation from Backlog Items
Tool Demonstrations
Mahja Sarschar Hochschule für Technik und Wirtschaft Berlin, Gefei Zhang HTW Berlin, Annika Nowak Capgemini
11:45
15m
Talk
Contextualized Data-Wrangling Code Generation in Computational Notebooks
Research Papers
Junjie Huang The Chinese University of Hong Kong, Daya Guo Sun-yat Sen University, Chenglong Wang Microsoft Research, Jiazhen Gu Chinese University of Hong Kong, Shuai Lu Microsoft Research, Jeevana Priya Inala Microsoft Research, Cong Yan Microsoft Research, Jianfeng Gao Microsoft Research, Nan Duan Microsoft Research, Michael Lyu The Chinese University of Hong Kong