TCSE logo 
 Sigsoft logo
Sustainability badge

This program is tentative and subject to change.

Tue 29 Apr 2025 14:00 - 15:30 at Canada Hall 3 Poster Area - ACM Student Research Posters and Judging 3
Thu 1 May 2025 12:06 - 12:12 at 204 - ACM Student Research Presentations

The use of Large Language Models (LLMs) for code generation has emerged as a rapidly growing field, gaining substantial traction within software engineering. However, ensuring the reliability and accuracy of generated code requires robust evaluation frameworks. To address this gap, Carlos et al. introduced the SWE-bench dataset, which consists of 2,294 GitHub issues paired with their corresponding pull requests, collected from 12 prominent Python repositories. This dataset has become a key benchmark for evaluating code generation models, with resolution rates prominently featured on the SWE-bench leaderboard. Despite its widespread adoption, the dataset has yet to undergo a systematic reliability assessment. Motivated by this gap, we conducted the first empirical study aimed at evaluating the reliability of the SWE-Bench dataset to ensure it provides meaningful and realistic model evaluations. We centered our analysis on the highest-performing model reported on the leaderboard at the time of the study: SWE-Agent + GPT-4. A thorough investigation was conducted by comparing the model-generated patches with the corresponding pull requests from the dataset. Our findings revealed two key issues: (1) 32.67% of successful cases were influenced by solution leakage, and (2) 31.08% succeeded due to weak test cases. When these problematic instances were excluded, the resolution rate of SWE-Agent + GPT4 dropped from 12.47% to 3.97%.

This program is tentative and subject to change.

Tue 29 Apr

Displayed time zone: Eastern Time (US & Canada) change

14:00 - 15:30
ACM Student Research Posters and Judging 3SRC - ACM Student Research Competition at Canada Hall 3 Poster Area
14:00
90m
Talk
Revisiting SWE-Bench: On the Importance of Data Quality for LLM-based Code Models
SRC - ACM Student Research Competition
Reem Aleithan York University, Canada
14:00
90m
Talk
On the Fly Input Refinement for Code Language Models
SRC - ACM Student Research Competition
Ravishka Shemal Rathnasuriya University of Texas at Dallas
14:00
90m
Talk
Program Feature-based Fuzzing Benchmarking
SRC - ACM Student Research Competition
Miao Miao The University of Texas at Dallas
14:00
90m
Talk
On the Automation of Code Review Tasks Through Cross-Task Knowledge Distillation
SRC - ACM Student Research Competition
Oussama Ben Sghaier DIRO, Université de Montréal

Thu 1 May

Displayed time zone: Eastern Time (US & Canada) change

11:00 - 12:30
ACM Student Research PresentationsSRC - ACM Student Research Competition at 204

A subset of finalist ACM SRC students will give short presentations in this session. That decision about who will present will be made after the poster sessions, and this schedule will be updated, so don’t rely on the precise timing until just before the session..

They all also have posters in Canada Hall 3 Poster area, with judging to be on Tuesday.

Awards will be announced in the banquet on Thursday evening.

11:00
6m
Talk
Automatic Fuzz Drivers for JavaScript with Type Distributions
SRC - ACM Student Research Competition
Mayant Mukul University of British Columbia
11:06
6m
Talk
CASS: Context-Aware Slice Summarization for Debugging Regression Failures
SRC - ACM Student Research Competition
Sahar Badihi University of British Columbia, Canada
11:12
6m
Talk
Characterising Algorithm Debt in Machine and Deep Learning Systems
SRC - ACM Student Research Competition
Emmanuel Iko-Ojo Simon Australian National University
11:18
6m
Talk
Consistent Graph Model Generation with Large Language Models
SRC - ACM Student Research Competition
Boqi Chen McGill University
11:24
6m
Talk
Enhancing OSS Remediation with Patch Backporting
SRC - ACM Student Research Competition
Lyuye Zhang Nanyang Technological University
11:30
6m
Talk
Identifying Performance-Sensitive Configurations in Software Systems with LLM-Driven Agents
SRC - ACM Student Research Competition
Zehao Wang Concordia University
11:36
6m
Talk
Improving Formal Methods VisualizationsFormal Methods
SRC - ACM Student Research Competition
Avinash Palliyil Georgia Institute of Technology
11:42
6m
Talk
MUARF: Leveraging Multi-Agent Workflows for Automated Code Refactoring
SRC - ACM Student Research Competition
Yisen Xu Software PErformance, Analysis, and Reliability (SPEAR) lab, Concordia University, Montreal, Canada
11:48
6m
Talk
On the Automation of Code Review Tasks Through Cross-Task Knowledge Distillation
SRC - ACM Student Research Competition
Oussama Ben Sghaier DIRO, Université de Montréal
11:54
6m
Talk
On the Fly Input Refinement for Code Language Models
SRC - ACM Student Research Competition
Ravishka Shemal Rathnasuriya University of Texas at Dallas
12:00
6m
Talk
Program Feature-based Fuzzing Benchmarking
SRC - ACM Student Research Competition
Miao Miao The University of Texas at Dallas
12:06
6m
Talk
Revisiting SWE-Bench: On the Importance of Data Quality for LLM-based Code Models
SRC - ACM Student Research Competition
Reem Aleithan York University, Canada
12:12
6m
Talk
The Balancing Act of Policies in Developing Machine Learning Explanations
SRC - ACM Student Research Competition
Jacob Tjaden Colby College
12:18
6m
Talk
To Mock or Not to Mock: Divergence in Mocking Practices Between LLM and Developers
SRC - ACM Student Research Competition
Hanbin Qin Stevens Institute of Technology
12:24
6m
Talk
Towards Compatibly Mitigating Technical Lag in Maven Projects
SRC - ACM Student Research Competition
Rui Lu East China Normal University
:
:
:
: