ICSE 2024
Fri 12 - Sun 21 April 2024 Lisbon, Portugal

Benchmark plays a pivotal role in advancing the research on the programming related tasks. In this study, we introduce, PyP4LLMSec, a Python benchmark designed to assess the security aspect of Python code generated by large language models (LLMs). Our methodology involves an analysis of Common Vulnerabilities and Exposures (CVEs) over the past two years. We identified 257 vulnerability-related commits associated with these CVEs across 143 open-source Python projects on GitHub. Subsequently, we conducted manual inspections of the vulnerable code, identifying and analyzing 295 code patches addressing vulnerabilities to generate Python code prompts at the file, class, and function granularity levels. As a result, we generated 2142 prompts with three distinct types of endings at various granularity levels, covering 15 different Common Weakness Enumeration (CWE) categories. To the best of our knowledge, this dataset represents the first collection of Python programming language prompts for scrutinizing the security of code generated by LLMs across different granularity levels. Our dataset, PyP4LLMSec, is publicly accessible on GitHub.

Sat 20 Apr

Displayed time zone: Lisbon change

14:00 - 15:30
Session 3: Keynote 2 + Position PapersLLM4Code at Luis de Freitas Branco
Chair(s): Lingming Zhang University of Illinois at Urbana-Champaign
14:00
50m
Keynote
Open development of Large Language Models for code with BigCode and StarCoder2
LLM4Code
Loubna Ben Allal Hugging Face
14:50
8m
Talk
Benchmarking the Security Aspect of Large Language Model-Based Code Generation
LLM4Code
Cheng Cheng Concordia University, Jinqiu Yang Concordia University
Pre-print
14:58
8m
Talk
Enhancing LLM-Based Coding Tools through Native Integration of IDE-Derived Static Context
LLM4Code
Yichen LI The Chinese University of Hong Kong, Yun Peng The Chinese University of Hong Kong, Yintong Huo The Chinese University of Hong Kong, Michael Lyu The Chinese University of Hong Kong
Pre-print
15:06
8m
Talk
Evaluating Fault Localization and Program Repair Capabilities of Existing Closed-Source General-Purpose LLMs
LLM4Code
Shengbei Jiang Beijing Jiaotong University, Jiabao Zhang Beijing Jiaotong University, Wei Chen Beijing Jiaotong University, Bo Wang Beijing Jiaotong University, Jianyi Zhou Huawei Cloud Computing Technologies Co., Ltd., Jie M. Zhang King's College London
Pre-print
15:14
8m
Talk
MoonBit: Explore the Design of an AI-Friendly Programming Language
LLM4Code
Haoxiang Fei International Digital Economy Academy, Yu Zhang International Digital Economy Academy, Hongbo Zhang International Digital Economy Academy, Yanlin Wang Sun Yat-sen University, Qing Liu International Digital Economy Academy
Pre-print
15:22
8m
Talk
Toward a New Era of Rapid Development: Assessing GPT-4-Vision's Capabilities in UML-Based Code Generation
LLM4Code
Gabor Antal University of Szeged, Richárd Vozár Department of Software Engineering, University of Szeged, Hungary, Rudolf Ferenc University of Szeged