ICST 2025
Mon 31 March - Fri 4 April 2025 Naples, Italy

This program is tentative and subject to change.

Thu 3 Apr 2025 11:00 - 11:15 at Room A - LLMs in Testing Chair(s): Valerio Terragni

Search-based test generators are effective at producing unit tests with high coverage. However, such automatically generated tests have no meaningful test and variable names, making them hard to understand and interpret by developers. On the other hand, large language models (LLMs) can generate highly readable test cases, but they are not able to match the effectiveness of search-based generators, in terms of achieved code coverage.

In this paper, we propose to combine the effectiveness of search-based generators with the readability of LLM generated tests. Our approach focuses on improving test and variable names produced by search-based tools, while keeping their semantics (i.e., their coverage) unchanged.

Our evaluation on nine industrial and open source LLMs show that our readability improvement transformations are overall semantically-preserving and stable across multiple repetitions. Moreover, a human study with ten professional developers, show that our LLM-improved tests are as readable as developer-written tests, regardless of the LLM employed.

This program is tentative and subject to change.

Thu 3 Apr

Displayed time zone: Amsterdam, Berlin, Bern, Rome, Stockholm, Vienna change

11:00 - 12:22
LLMs in TestingResearch Papers / Short Papers, Vision and Emerging Results at Room A
Chair(s): Valerio Terragni University of Auckland
11:00
15m
Talk
Improving the Readability of Automatically Generated Tests using Large Language Models
Research Papers
Matteo Biagiola Università della Svizzera italiana, Gianluca Ghislotti Università della Svizzera italiana, Paolo Tonella USI Lugano
11:15
15m
Talk
Test Wars: A Comparative Study of SBST, Symbolic Execution, and LLM-Based Approaches to Unit Test Generation
Research Papers
Azat Abdullin JetBrains Research, TU Delft, Pouria Derakhshanfar JetBrains Research, Annibale Panichella Delft University of Technology
11:30
15m
Talk
Benchmarking Open-source Large Language Models For Log Level Suggestion
Research Papers
Yi Wen HENG Concordia University, Zeyang Ma Concordia University, Zhenhao Li York University, Dong Jae Kim DePaul University, Tse-Hsun (Peter) Chen Concordia University
11:45
15m
Talk
Understanding and Enhancing Attribute Prioritization in Fixing Web UI Tests with LLMs
Research Papers
Zhuolin Xu Concordia University, Qiushi Li Concordia University, Shin Hwei Tan Concordia University
12:00
15m
Talk
Benchmarking Generative AI Models for Deep Learning Test Input Generation
Research Papers
Maryam Maryam , Matteo Biagiola Università della Svizzera italiana, Andrea Stocco Technical University of Munich, fortiss, Vincenzo Riccio University of Udine
Pre-print
12:15
7m
Talk
Leveraging Large Language Models for Explicit Wait Management in End-to-End Web Testing
Short Papers, Vision and Emerging Results
Dario Olianas DIBRIS, University of Genova, Italy, Maurizio Leotta DIBRIS, University of Genova, Italy, Filippo Ricca Università di Genova
:
:
:
: