ICSE 2026
Sun 12 - Sat 18 April 2026 Rio de Janeiro, Brazil

This program is tentative and subject to change.

Fri 17 Apr 2026 15:15 - 15:30 at Asia I - AI for Software Engineering 23

Large Language Models (LLMs) have shown great potential in Automated Program Repair (APR). Test inputs, being crucial for reasoning the root cause of failures, are always included in the prompt for LLM-based APR. Unfortunately, LLMs struggle to retain key information in long prompts. When the test inputs are extensive in the prompt, this may trigger the “lost-in-the-middle” issue, compromising repair performance. ReduceFix prompts an LLM to generate a reducer that minimizes failure-inducing test inputs without human effort, and then feeds the reduced failure-inducing inputs to guide patch generation.

For targeted evaluation, we constructed LFTBench, the first long-input APR benchmark with 200 real bugs from 20 programming tasks, each paired with a failure‑inducing input whose median size is 1 MB. On this benchmark, ReduceFix shrinks inputs by 89.1% on average and improves overall pass@10 by up to 53.8% relative to a prompt that includes the original test, and by 17.6% compared with omitting the test entirely. Adding the same reduction step to ChatRepair and CREF increases their fix rate by 21.3% and 2.6%, respectively, without other changes. Our gains hold against a ddmin‑only reducing template baseline and transfer to repository‑level OSS‑Fuzz cases. Ablation studies further highlight the impact of input length and compressed failure information on repair success. These results underscore that automatically reducing failing inputs is a practical and powerful complement to LLM-based APR, significantly improving its scalability and effectiveness.

This program is tentative and subject to change.

Fri 17 Apr

Displayed time zone: Brasilia, Distrito Federal, Brazil change

14:00 - 15:30
AI for Software Engineering 23Research Track / Demonstrations / Journal-first Papers at Asia I
14:00
15m
Talk
CI-Bench: A Framework for Evaluating Large Language Model Tools on CI Failures
Demonstrations
Raian Latif Nabil University of California, Davis, Hao-Nan Zhu University of California, Davis, Cindy Rubio-González University of California at Davis
14:15
15m
Talk
Assessing the Latent Automated Program Repair Capabilities of Large Language Models using Round-Trip Translation
Journal-first Papers
Fernando Vallecillos Ruiz Simula Research Laboratory, Anastasiia Grishina Simula Research Laboratory, Max Hort Simula Research Laboratory, Leon Moonen Simula Research Laboratory
14:30
15m
Talk
XRFix: Exploring Performance Bug Repair of Extended Reality Applications with Large Language Models
Research Track
Jingwen Wu Department of Computer Science, Hong Kong Baptist University, Hanyang Guo School of Software Engineering, Sun Yat-sen University, Hong-Ning Dai Department of Computer Science, Hong Kong Baptist University, Xiapu Luo Hong Kong Polytechnic University
DOI Pre-print
14:45
15m
Talk
Synthetic Repo-level Bug Dataset for Training Automated Program Repair Models
Research Track
Minh V. T. Pham FPT Software AI Center, Huy N. Phan FPT Software AI Center, Hoang Nhat Phan Nanyang Technological University, Cuong Chi Le The University of Texas at Dallas, Tien N. Nguyen University of Texas at Dallas, Nghi D. Q. Bui Google Research
15:00
15m
Talk
PredicateFix: Repairing Static Analysis Alerts with Bridging Predicates
Research Track
Yuan-An Xiao Peking University, Weixuan Wang Peking University, Dong Liu Center Research Institute, ZTE Coporation, China, Junwei Zhou Center Research Institute, ZTE Coporation, China, Shengyu Cheng ZTE Corporation, Yingfei Xiong Peking University
Pre-print
15:15
15m
Talk
Input Reduction Enhanced LLM-based Program Repair
Research Track
Boyang Yang Yanshan University, Luyao Ren Peking University, Xin Yin Zhejiang University, Jiadong Ren Yanshan University, Haoye Tian Aalto University, Shunfu Jin Yanshan University
DOI Pre-print