Software failures inform engineering work, standards, regulations. For example, the Log4J vulnerability brought government and industry attention to evaluating and securing software supply chains. Retrospective failure analysis is thus a valuable line of software engineering research. Accessing private engineering records is difficult, so such analyses tend to use information reported by the news media. However, prior works in this direction have relied on manual analysis. That has limited the scale of their analyses. The community lacks automated support to enable such analyses to consider a wide range of news sources and incidents.
In this paper, we propose the Failure Analysis Investigation with LLMs (FAIL) system to fill this gap. FAIL collects, analyzes, and summarizes software failures as reported in the news. FAIL groups articles that describe the same incidents. It then analyzes incidents using existing taxonomies for postmortems, faults, and system characteristics. To tune and evaluate FAIL, we followed the methods of prior works by manually analyzing 31 software failures. FAIL achieved an F1 score of 90% for collecting news about software failures, a V-measure of 0.98 for merging articles reporting on the same incident, and extracted 90% of the facts about failures. We then applied FAIL to a total of 137,427 news articles from 11 providers published between 2010 and 2022. FAIL identified and analyzed 2,457 distinct failures reported across 4,184 articles. Our findings include: (1) current generation of large language models are capable of identifying news articles that describe failures, and analyzing them according to structured taxonomies; (2) high recurrences of similar failures within organizations and across organizations; and (3) severity of the consequences of software failures have increased over the past decade. The full FAIL database is available so that researchers, engineers, and policymakers can learn from a diversity of software failures.
Wed 30 OctDisplayed time zone: Pacific Time (US & Canada) change
10:30 - 12:00 | Log and trace; failure and faultResearch Papers / Industry Showcase at Carr Chair(s): Yiming Tang Rochester Institute of Technology | ||
10:30 15mTalk | Demonstration-Free: Towards More Practical Log Parsing with Large Language Models Research Papers | ||
10:45 15mTalk | Unlocking the Power of Numbers: Log Compression via Numeric Token Parsing Research Papers | ||
11:00 15mTalk | Towards Synthetic Trace Generation of Modeling Operations using In-Context Learning Approach Research Papers Vittoriano Muttillo University of Teramo, Claudio Di Sipio University of l'Aquila, Riccardo Rubei University of L'Aquila, Luca Berardinelli Johannes Kepler University Linz, MohammadHadi Dehghani Johannes Kepler University Linz | ||
11:15 15mTalk | DeployFix: Dynamic Repair of Software Deployment Failures via Constraint Solving Industry Showcase Haoyu Liao East China Normal University, Jianmei Guo East China Normal University, Bo Huang East China Normal University, Yujie Han East China Normal University, Dingyu Yang Zhejiang University, Kai Shi Alibaba Group, Jonathan Ding Intel, Guoyao Xu Alibaba Group, Guodong Yang Alibaba Group, Liping Zhang Alibaba Group | ||
11:30 15mTalk | FAIL: Analyzing Software Failures from the News Using LLMs Research Papers Dharun Anandayuvaraj Purdue University, Matthew Campbell Purdue University, Arav Tewari Purdue University, James C. Davis Purdue University DOI Pre-print | ||
11:45 15mTalk | Do not neglect what's on your hands: localizing software faults with exception trigger stream Research Papers Xihao Zhang School of Computer Science, Wuhan University, Yi Song School of Computer Science, Wuhan University, Xiaoyuan Xie Wuhan University, Qi Xin Wuhan University, Chenliang Xing School of Computer Science, Wuhan University |