RepoMasterEval: Evaluating Code Completion via Real-World Repositories
This program is tentative and subject to change.
With the growing reliance on automated code completion tools in software development, the need for comprehensive evaluation benchmarks has become critical. Existing benchmarks focus more on code completion in function and class level by providing text descriptions to prompt the model. By contrast, such descriptive prompt is commonly unavailable in real development and code completion can occur in wider range of situations such as in the middle of a function or a code block. These limitations makes existing evaluation benchmarks poorly align with the practical scenarios of code completion tools. In this paper, we propose RepoMasterEval, a novel benchmark for evaluating code completion models constructed from real-world repositories. Each benchmark datum is generated by masking a code snippet (ground truth) from one source code file with existing test suites. To improve test accuracy of model generated code, we employ mutation testing to measure the effectiveness of the test cases and we manually crafted new test cases for those test suites with low mutation score. Our empirical evaluation on 10 state-of-the-art models shows that test argumentation is critical in improving the accuracy of the benchmark and RepoMasterEval is able to report variance in model performance in real-world scenarios. The deployment of RepoMasterEval also revealed that the benchmark is useful to give accurate feedback during model training and the score is in high correlation with the model’s performance in practice.
This program is tentative and subject to change.
Tue 18 NovDisplayed time zone: Seoul change
16:00 - 17:00 | |||
16:00 10mTalk | Automated Prompt Generation for Code Intelligence: An Empirical study and Experience in WeChat Industry Showcase Kexing Ji , Shiyun Fu The Chinese University of Hong Kong, Cuiyun Gao Harbin Institute of Technology, Shenzhen, Yujia Chen The Chinese University of Hong Kong, Zezhou Yang Tencent Inc., Chaozheng Wang The Chinese University of Hong Kong, Yuetang Deng Tencent | ||
16:10 10mTalk | Evaluating Large Language Models for Functional and Maintainable Code in Industrial Settings: A Case Study at ASML Industry Showcase Yash Mundhra Delft University of Technology, Max Valk ASML, Maliheh Izadi Delft University of Technology | ||
16:20 10mTalk | IntelliTopo: An IaC Generation Service for Industrial Network Topology Construction Industry Showcase Mingyu Shao Harbin Institute of Technology, Shenzhen; PengCheng Laboratory, Zhao Liu PengCheng Laboratory, Weihong Han Peng Cheng Laboratory, Cuiyun Gao Harbin Institute of Technology, Shenzhen, Jiachen Liu Harbin Institute of Technology, Shenzhen, Qing Liao Harbin Institute of Technology | ||
16:30 10mTalk | RepoMasterEval: Evaluating Code Completion via Real-World Repositories Industry Showcase Qinyun Wu Bytedance Ltd., Chao Peng ByteDance, Pengfei Gao ByteDance, Ruida Hu Harbin Institute of Technology, Shenzhen, Haoyu Gan ByteDance, Bo Jiang Bytedance Network Technology, Jinhe Tang ByteDance, Zhiwen Deng ByteDance, Zhanming Guan ByteDance, Cuiyun Gao Harbin Institute of Technology, Shenzhen, Xia Liu ByteDance, Ping Yang Bytedance Network Technology | ||
16:40 10mTalk | Multiple Schema-Conformant Declarative Code Generation NIER Track | ||
16:50 10mTalk | Tuning LLM-based Code Optimization via Meta-Prompting: An Industrial Perspective Industry Showcase Jingzhi Gong University of Leeds, Rafail Giavrimis Turing Intelligence Technology, Paul Brookes TurinTech AI, Vardan Voskanyan TurinTech AI, Fan Wu TurinTech AI, Mari Ashiga University of West London/TurinTech AI, Matthew Truscott TurinTech AI, Michail Basios Turing Intelligence Technology, Leslie Kanthan TurinTech AI, Jie Xu University of Leeds, Zheng Wang University of Leeds |