ASE 2025
Sun 16 - Thu 20 November 2025 Seoul, South Korea

This program is tentative and subject to change.

Tue 18 Nov 2025 11:20 - 11:30 at Grand Hall 1 - Code Generation 2

Trustworthy evaluation methods for code snippets play a crucial role in neural code generation. Traditional methods, which either rely on reference solutions or require executable test cases, have inherent limitation in flexibility and scalability. The recent LLM-as-Judge methodology offers a promising alternative by directly evaluating functional consistency between the problem description and the generated code. To systematically understand the landscape of these LLM-as-Judge methods, we conduct a comprehensive empirical study across three diverse datasets. Our investigation reveals the pros and cons of two categories of LLM-as-Judge methods: the methods based on general foundation models can achieve good performance but require complex prompts and lack explainability, while the methods based on reasoning foundation models provide better explainability with simpler prompts but demand substantial computational resources due to their large parameter sizes.

To address these limitations, we propose CODE-DITING, a novel code evaluation method that balances accuracy, efficiency and explainability. We develop a data distillation framework that effectively transfers reasoning capabilities from DeepSeek-R1671B to our CODE-DITING 1.5B and 7B models, significantly enhancing evaluation explainability and reducing the computational cost. With the majority vote strategy in the inference process, CODE-DITING 1.5B outperforms all models with the same magnitude of parameters and achieves performance which would normally exhibit in a model with 5 times of parameter scale. CODE-DITING 7B surpasses GPT-4o and DeepSeek-V3 671B, even though it only uses 1% of the parameter volume of these large models. Further experiments show that CODEDITING is robust to preference leakage and can serve as a promising alternative for code evaluation.

This program is tentative and subject to change.

Tue 18 Nov

Displayed time zone: Seoul change

11:00 - 12:30
11:00
10m
Talk
Coverage-Based Harmfulness Testing for LLM Code Transformation
Research Papers
Honghao Tan Concordia University, Haibo Wang Concordia University, Diany Pressato Concordia University, Yisen Xu Software PErformance, Analysis, and Reliability (SPEAR) lab, Concordia University, Montreal, Canada, Shin Hwei Tan Concordia University
11:10
10m
Talk
RealisticCodeBench: Towards More Realistic Evaluation of Large Language Models for Code Generation
Research Papers
Xiao Yu Zhejiang University, Haoxuan Chen Wuhan University of Technology, Lei Liu Xi’an Jiaotong University, Xing Hu Zhejiang University, Jacky Keung City University of Hong Kong, Xin Xia Zhejiang University
11:20
10m
Talk
Code-DiTing: Automatic Evaluation of Code Generation without References or Test Cases
Research Papers
Guang Yang , Yu Zhou Nanjing University of Aeronautics and Astronautics, Xiang Chen Nantong University, Wei Zheng Northwestern Polytechnical University, Xing Hu Zhejiang University, Xin Zhou Singapore Management University, Singapore, David Lo Singapore Management University, Taolue Chen Birkbeck, University of London
Pre-print
11:30
10m
Talk
An Agent-based Evaluation Framework for Complex Code Generation
Research Papers
Xinchen Wang Harbin Institute of Technology, Pengfei Gao ByteDance, Chao Peng ByteDance, Ruida Hu Harbin Institute of Technology, Shenzhen, Cuiyun Gao Harbin Institute of Technology, Shenzhen
11:40
10m
Talk
PseudoFix: Refactoring Distorted Structures in Decompiled C Pseudocode
Research Papers
Gangyang Li University of Science and Technology of China, Xiuwei Shang University of Science and Technology of China, Shaoyin Cheng University of Science and Technology of China, junqi zhang University of Science and Technology of China, Li Hu , Xu Zhu University of Science and Technology of China, Weiming Zhang University of Science and Technology of China, Nenghai Yu School of Cyber Security, University of Science and Technology of China
11:50
10m
Talk
Evaluating and Improving Framework-based Parallel Code Completion with Large Language Models
Research Papers
Ke Liu , Qinglin Wang Shandong Normal University, Xiang Chen Nantong University, Guang Yang , YiGui Feng National University of Defense Technology, Gencheng Liu National University of Defense Technology, Jie Liu Institute of Software, Chinese Academy of Sciences
12:00
10m
Talk
Variational Prefix Tuning for diverse and accurate code summarization using pre-trained language models
Journal-First Track
Junda Zhao Department of Mechanical and Industrial Engineering, University of Toronto, Yuliang Song Department of Mechanical and Industrial Engineering, University of Toronto, Eldan Cohen Department of Mechanical and Industrial Engineering, University of Toronto
12:10
10m
Talk
Effective Code Membership Inference for Code Completion Models via Adversarial Prompts
Research Papers
Yuan Jiang Harbin Institute of Technology, Zehao Li Harbin Institute of Technology, Shan Huang East China Normal University, Christoph Treude Singapore Management University, Xiaohong Su Harbin Institute of Technology, Tiantian Wang Harbin Institute of Technology
12:20
10m
Talk
LongCodeZip: Compress Long Context for Code Language Models
Research Papers
Yuling Shi Shanghai Jiao Tong University, Yichun Qian Stanford University, Hongyu Zhang Chongqing University, Beijun Shen Shanghai Jiao Tong University, Xiaodong Gu Shanghai Jiao Tong University
Pre-print Media Attached