Write a Blog >>
ICSE 2023
Sun 14 - Sat 20 May 2023 Melbourne, Australia
Fri 19 May 2023 16:45 - 16:52 at Level G - Plenary Room 1 - Software quality Chair(s): Valentina Lenarduzzi

Static analysis tools find defects in code, checking code against rules to reveal potential defects. Many studies have evaluated these tools by measuring their ability to detect known defects in code. But these studies measure the current state of tools rather than their future potential to find more defects. To investigate the prospects for tools to find more defects, we conducted a study where we formulated each issue raised by a code reviewer as a violation of a rule, which we then compared to what static analysis tools might potentially check. We first gathered a corpus of 1323 defects found through code review. Through a qualitative analysis process, for each defect we identified a violated rule and the type of Static Analysis Tool (SAT) which might check this rule. We found that SATs might, in principle, be used to detect as many as 76% of code review defects, considerably more than current tools have been demonstrated to successfully detect. Among a variety of types of SATs, Style Checkers and AST Pattern Checkers had the broadest coverage of defects, each with the potential to detect 25% of all code review defects. We found that static analysis tools might be able to detect more code review defects by better supporting the creation of project-specific rules. We also investigated the characteristics of code review defects not detectable by traditional static analysis techniques, which to detect might require tools which simulate human judgements about code.

Fri 19 May

Displayed time zone: Hobart change

15:45 - 17:15
15:45
15m
Talk
DuetCS: Code Style Transfer through Generation and Retrieval
Technical Track
Binger Chen Technische Universität Berlin, Ziawasch Abedjan Leibniz Universität Hannover
16:00
15m
Talk
Understanding Why and Predicting When Developers Adhere to Code-Quality Standards
SEIP - Software Engineering in Practice
Manish Motwani Georgia Institute of Technology, Yuriy Brun University of Massachusetts
Pre-print
16:15
15m
Talk
Code Compliance Assessment as a Learning Problem
SEIP - Software Engineering in Practice
16:30
15m
Talk
An Empirical Study on Quality Issues of Deep Learning Platform
SEIP - Software Engineering in Practice
Yanjie Gao Microsoft Research, Xiaoxiang Shi , Haoxiang Lin Microsoft Research, Hongyu Zhang The University of Newcastle, Hao Wu , Rui Li , Mao Yang Microsoft Research
Pre-print
16:45
7m
Talk
Can static analysis tools find more defects? A qualitative study of design rule violations found by code review
Journal-First Papers
Sahar Mehrpour George Mason University, USA, Thomas LaToza George Mason University
16:52
7m
Talk
DebtFree: minimizing labeling cost in self-admitted technical debt identification using semi-supervised learning
Journal-First Papers
Huy Tu North Carolina State University, USA, Tim Menzies North Carolina State University
Link to publication Pre-print
17:00
7m
Talk
FIXME: synchronize with database! An empirical study of data access self-admitted technical debt
Journal-First Papers
Biruk Asmare Muse Polytechnique Montréal, Csaba Nagy Software Institute - USI, Lugano, Anthony Cleve University of Namur, Foutse Khomh Polytechnique Montréal, Giuliano Antoniol Polytechnique Montréal
17:07
7m
Talk
How does quality deviate in stable releases by backporting?
NIER - New Ideas and Emerging Results
Jarin Tasnim University of Saskatchewan, Debasish Chakroborti University of Saskatchewan, Chanchal K. Roy University of Saskatchewan, Kevin Schneider University of Saskatchewan
Link to publication Pre-print