Revisiting SWE-Bench: On the Importance of Data Quality for LLM-based Code Models
This program is tentative and subject to change.
Thu 1 May 2025 12:06 - 12:12 at 204 - ACM Student Research Presentations
The use of Large Language Models (LLMs) for code generation has emerged as a rapidly growing field, gaining substantial traction within software engineering. However, ensuring the reliability and accuracy of generated code requires robust evaluation frameworks. To address this gap, Carlos et al. introduced the SWE-bench dataset, which consists of 2,294 GitHub issues paired with their corresponding pull requests, collected from 12 prominent Python repositories. This dataset has become a key benchmark for evaluating code generation models, with resolution rates prominently featured on the SWE-bench leaderboard. Despite its widespread adoption, the dataset has yet to undergo a systematic reliability assessment. Motivated by this gap, we conducted the first empirical study aimed at evaluating the reliability of the SWE-Bench dataset to ensure it provides meaningful and realistic model evaluations. We centered our analysis on the highest-performing model reported on the leaderboard at the time of the study: SWE-Agent + GPT-4. A thorough investigation was conducted by comparing the model-generated patches with the corresponding pull requests from the dataset. Our findings revealed two key issues: (1) 32.67% of successful cases were influenced by solution leakage, and (2) 31.08% succeeded due to weak test cases. When these problematic instances were excluded, the resolution rate of SWE-Agent + GPT4 dropped from 12.47% to 3.97%.
This program is tentative and subject to change.
Tue 29 AprDisplayed time zone: Eastern Time (US & Canada) change
14:00 - 15:30 | ACM Student Research Posters and Judging 3SRC - ACM Student Research Competition at Canada Hall 3 Poster Area | ||
14:00 90mTalk | Revisiting SWE-Bench: On the Importance of Data Quality for LLM-based Code Models SRC - ACM Student Research Competition Reem Aleithan York University, Canada | ||
14:00 90mTalk | On the Fly Input Refinement for Code Language Models SRC - ACM Student Research Competition Ravishka Shemal Rathnasuriya University of Texas at Dallas | ||
14:00 90mTalk | Program Feature-based Fuzzing Benchmarking SRC - ACM Student Research Competition Miao Miao The University of Texas at Dallas | ||
14:00 90mTalk | On the Automation of Code Review Tasks Through Cross-Task Knowledge Distillation SRC - ACM Student Research Competition Oussama Ben Sghaier DIRO, Université de Montréal |
Thu 1 MayDisplayed time zone: Eastern Time (US & Canada) change
11:00 - 12:30 | ACM Student Research PresentationsSRC - ACM Student Research Competition at 204 A subset of finalist ACM SRC students will give short presentations in this session. That decision about who will present will be made after the poster sessions, and this schedule will be updated, so don’t rely on the precise timing until just before the session.. They all also have posters in Canada Hall 3 Poster area, with judging to be on Tuesday. Awards will be announced in the banquet on Thursday evening. | ||
11:00 6mTalk | Automatic Fuzz Drivers for JavaScript with Type Distributions SRC - ACM Student Research Competition Mayant Mukul University of British Columbia | ||
11:06 6mTalk | CASS: Context-Aware Slice Summarization for Debugging Regression Failures SRC - ACM Student Research Competition Sahar Badihi University of British Columbia, Canada | ||
11:12 6mTalk | Characterising Algorithm Debt in Machine and Deep Learning Systems SRC - ACM Student Research Competition Emmanuel Iko-Ojo Simon Australian National University | ||
11:18 6mTalk | Consistent Graph Model Generation with Large Language Models SRC - ACM Student Research Competition Boqi Chen McGill University | ||
11:24 6mTalk | Enhancing OSS Remediation with Patch Backporting SRC - ACM Student Research Competition Lyuye Zhang Nanyang Technological University | ||
11:30 6mTalk | Identifying Performance-Sensitive Configurations in Software Systems with LLM-Driven Agents SRC - ACM Student Research Competition Zehao Wang Concordia University | ||
11:36 6mTalk | Improving Formal Methods VisualizationsFormal Methods SRC - ACM Student Research Competition Avinash Palliyil Georgia Institute of Technology | ||
11:42 6mTalk | MUARF: Leveraging Multi-Agent Workflows for Automated Code Refactoring SRC - ACM Student Research Competition Yisen Xu Software PErformance, Analysis, and Reliability (SPEAR) lab, Concordia University, Montreal, Canada | ||
11:48 6mTalk | On the Automation of Code Review Tasks Through Cross-Task Knowledge Distillation SRC - ACM Student Research Competition Oussama Ben Sghaier DIRO, Université de Montréal | ||
11:54 6mTalk | On the Fly Input Refinement for Code Language Models SRC - ACM Student Research Competition Ravishka Shemal Rathnasuriya University of Texas at Dallas | ||
12:00 6mTalk | Program Feature-based Fuzzing Benchmarking SRC - ACM Student Research Competition Miao Miao The University of Texas at Dallas | ||
12:06 6mTalk | Revisiting SWE-Bench: On the Importance of Data Quality for LLM-based Code Models SRC - ACM Student Research Competition Reem Aleithan York University, Canada | ||
12:12 6mTalk | The Balancing Act of Policies in Developing Machine Learning Explanations SRC - ACM Student Research Competition Jacob Tjaden Colby College | ||
12:18 6mTalk | To Mock or Not to Mock: Divergence in Mocking Practices Between LLM and Developers SRC - ACM Student Research Competition Hanbin Qin Stevens Institute of Technology | ||
12:24 6mTalk | Towards Compatibly Mitigating Technical Lag in Maven Projects SRC - ACM Student Research Competition Rui Lu East China Normal University |