Code Ranking with Human-Inspired Agent-Based Framework
Pre-trained large language models (LLMs) have emerged as a breakthrough technology in code intelligence such as code generation. Recently, many works have found that LLMs can generate a correct code solution when it is allowed to make numerous attempts. Consequently, a recent trend is to do a large-scale sampling of codes from LLMs and then rank the code to select the most suitable code (A process called code ranking). A common code ranking approach ranks the code by running it against a set of LLM-generated test cases in the form of assert statements. However, existing approaches overlook how humans design test cases through systematic behaviors, which are essential for creating reliable tests for code ranking. Moreover, humans often use input–output examples to clarify and articulate the intended functionality, rather than merely to rank code. To address these gaps, we propose RankAgent, a human-inspired agent-based framework that systematically simulates human test design behaviors. Extensive experiments on five LLMs (including both open- and closed-source models) and two benchmarks (HumanEval+ and LiveCodeBench) show that RankAgent achieves notable and consistent improvements.
Mon 13 AprDisplayed time zone: Brasilia, Distrito Federal, Brazil change
16:00 - 18:00 | Session 7 - LLM-Based Agents for Software Engineering TasksJournal First / Replications and Negative Results (RENE) / Research Track / ICPC Program at Europa II Chair(s): Wesley K.G. Assunção North Carolina State University, Banani Roy University of Saskatchewan | ||
16:00 10mTalk | LLMs for Qualitative Data Analysis Fail on Security-specific Comments in Human Experiments Replications and Negative Results (RENE) Maria Camporese University of Trento, Fabio Massacci University of Trento; Vrije Universiteit Amsterdam, Yuanjun Gong University of Trento Pre-print File Attached | ||
16:10 10mTalk | Do comments and expertise still matter? An experiment on programmers’ adoption of AI-generated JavaScript code Journal First Changwen LI , Christoph Treude Singapore Management University, Ofir Turel The University of Melbourne | ||
16:20 10mTalk | Reducing Token Usage of State-in-Context Agents using Minification Replications and Negative Results (RENE) | ||
16:30 10mTalk | Agile Story-Point Estimation: Is RAG a Better Way to Go? Replications and Negative Results (RENE) Lamyea Maha University of Saskatchewan, Tajmilur Rahman Gannon University, Chanchal K. Roy University of Saskatchewan DOI Pre-print | ||
16:40 10mTalk | Improved Bug Localization with AI Agents Leveraging Hypothesis and Dynamic Cognition Research Track Pre-print Media Attached | ||
16:50 10mTalk | Code Ranking with Human-Inspired Agent-Based Framework Research Track Liuwen Cao South China University of Technology, liang jiaxi , Jiexin Wang South China University of Technology, Yi Cai School of Software Engineering, South China University of Technology, Guangzhou, China | ||
17:00 20mLive Q&A | Joint QA and Discussion ICPC Program | ||
17:20 40mAwards | ICPC Awards and Closing Session ICPC Program | ||