Developers often run tests to check that their latest changes to a code repository did not break any previously working functionality. Ideally, any new test failures would indicate regressions caused by the latest changes. However, some test failures may not be due to the latest changes but due to non-determinism in the tests, popularly called flaky tests. The typical way to detect flaky tests is to rerun failing tests repeatedly. Unfortunately, rerunning failing tests can be costly and can slow down the development cycle.
We present the first extensive evaluation of rerunning failing tests and propose a new technique, called DeFlaker, that detects if a test failure is due to a flaky test without rerunning and with very low runtime overhead. DeFlaker monitors the coverage of latest code changes and marks as flaky any newly failing test that did not execute any of the changes. We deployed DeFlaker live, in the build process of 96 Java projects on TravisCI, and found 87 previously unknown flaky tests in 10 of these projects. We also ran experiments on project histories, where DeFlaker detected 1,874 flaky tests from 4,846 failures, with a low false alarm rate (1.5%). DeFlaker had a higher recall (95.5% vs. 23%) of confirmed flaky tests than Maven’s default flaky test detector.
Thu 31 MayDisplayed time zone: Amsterdam, Berlin, Bern, Rome, Stockholm, Vienna change
11:00 - 12:30 | |||
11:00 20mTalk | DeFlaker: Automatically Detecting Flaky Tests Technical Papers Jonathan Bell George Mason University, Owolabi Legunsen University of Illinois at Urbana-Champaign, Michael Hilton Carnegie Mellon University, USA, Lamyaa Eloussi , Tifany Yung , Darko Marinov University of Illinois at Urbana-Champaign Pre-print Media Attached | ||
11:20 20mTalk | DetReduce: Minimizing Android GUI Test Suites for Regression Testing Technical Papers Wontae Choi , Koushik Sen University of California, Berkeley, George Necula University of California, Berkeley, Wenyu Wang University of Illinois at Urbana-Champaign | ||
11:40 20mTalk | Time to Clean your Test Objectives Technical Papers Michaël Marcozzi Imperial College London, Sébastien Bardin , Nikolai Kosmatov , Mike Papadakis University of Luxembourg, Virgile Prevosto , Loïc Correnson Link to publication DOI File Attached | ||
12:00 20mTalk | Prioritizing Browser Environments for Web Application Test Execution Technical Papers Junghyun Kwon , In-Young Ko Korea Advanced Institute of Science and Technology, Gregg Rothermel University of Nebraska - Lincoln | ||
12:20 10mTalk | Q&A in groups Technical Papers |