EASE 2025
Tue 17 - Fri 20 June 2025 Istanbul, Turkey
Wed 18 Jun 2025 16:35 - 16:45 at Senate Hall - LLMs for SE (Code) Chair(s): Stefan Wagner

Code obfuscation is the conversion of original source code into a functionally equivalent but less readable form, aiming to prevent reverse engineering and intellectual property theft. This is a challenging task since it is crucial to maintain functional correctness of the code while substantially disguising the input code. The recent development of large language models (LLMs) paves the way for practical applications in different domains, including software engineering. This work performs an empirical study on the ability of LLMs to obfuscate Python source code and introduces a metric (i.e., semantic elasticity) to measure the quality degree of obfuscated code. We experimented with 3 leading LLMs, i.e., Claude-3.5-Sonnet, Gemini-1.5, GPT-4-Turbo across 30 Python functions from diverse computational domains. Our findings reveal GPT-4-Turbo’s remarkable effectiveness with few-shot prompting (81% pass rate versus 29% standard prompting), significantly outperforming both Gemini-1.5 (39%) and Claude-3.5-Sonnet (30%). Notably, we discovered a counter-intuitive “obfuscation by simplification” phenomenon where models consistently reduce rather than increase cyclomatic complexity. This study provides a methodological framework for evaluating AI-driven obfuscation while highlighting promising directions for leveraging LLMs in software security.

Wed 18 Jun

Displayed time zone: Athens change

15:30 - 17:00
LLMs for SE (Code)Short Papers, Emerging Results / AI Models / Data / Research Papers at Senate Hall
Chair(s): Stefan Wagner Technical University of Munich
15:30
10m
Talk
Exploring Zero-Shot App Review Classification with ChatGPT: Challenges and Potential
Short Papers, Emerging Results
Mohit Chaudhary TCS Research, Chirag Jain TCS Research, Preethu Rose Anish TCS Research
Pre-print
15:40
10m
Talk
How Are We Doing With Using AI-Based Programming Assistants For Privacy-Related Code Generation? The Developers' Experience
Short Papers, Emerging Results
Kashumi Madampe Monash University, Australia, John Grundy Monash University, Nalin Arachchilage RMIT University
15:50
15m
Talk
Do Prompt Patterns Affects Code Quality? A First Empirical Assessment of ChatGPT Generated Code
Research Papers
Antonio Della Porta University of Salerno, Stefano Lambiase University of Salerno, Fabio Palomba University of Salerno
Pre-print
16:05
15m
Talk
On Simulation-Guided LLM-based Code Generation for Safe Autonomous Driving Software
AI Models / Data
Ali Nouri Volvo cars & Chalmers University of Technology, Johan Andersson Chalmers University of Technology, Kailash De Jesus Hornig Chalmers University of Technology, Zhnnan Fei Volvo Cars, Emil Knabe Volvo Cars, Hakan Sivencrona Volvo Cars, Beatriz Cabrero-Daniel University of Gothenburg, Christian Berger Chalmers University of Technology, Sweden
Pre-print
16:20
15m
Talk
An Empirical study on LLM-based Log Retrieval for Software Engineering Metadata Management
AI Models / Data
Simin Sun Chalmers University of Technology and University of Gothenburg, Yuchuan Jin Zenseact, Miroslaw Staron Chalmers University of Technology and University of Gothenburg
16:35
10m
Talk
Simplicity by Obfuscation: Evaluating LLM-Driven Code Transformation with Semantic Elasticity
Short Papers, Emerging Results
LORENZO DE TOMASI University of L’Aquila, Claudio Di Sipio University of l'Aquila, Antinisca Di Marco University of L'Aquila, Phuong T. Nguyen University of L’Aquila
Pre-print
16:45
15m
Paper
Small Models, Big Tasks: An Exploratory Empirical Study on Small Language Models for Function Calling
AI Models / Data
Ishan Kavathekar International Institute of Information Technology, Hyderabad, Raghav Donakanti International Institute of Information Technology, Hyderabad, Ponnurangam Kumaraguru International Institute of Information Technology, Hyderabad, Karthik Vaidhyanathan IIIT Hyderabad
Pre-print
:
:
:
: