ICSE 2024
Fri 12 - Sun 21 April 2024 Lisbon, Portugal

While deep learning (DL) has emerged as a powerful technique, its benefits must be carefully considered in relation to computational costs. Specifically, although DL methods have achieved strong performance in log anomaly detection, they often require extended time for log preprocessing and model training, hindering their adoption in distributed cloud systems that require rapid construction and inference for log anomaly detection.

This paper investigates the superiority of DL methods compared to simpler techniques in log anomaly detection. We evaluate basic algorithms (e.g., KNN, SLFN) and DL approaches (e.g., CNN, NeuralLog) on five public log anomaly detection datasets (e.g., HDFS). Our findings demonstrate that simple algorithms outperform DL methods in both time efficiency and accuracy. For instance, on the Thunderbird dataset, the K-nearest neighbor algorithm trains 1,000 times faster than NeuralLog while achieving a higher F1-Score by 0.0625. We also identify three factors contributing to this phenomenon, which are: (1) redundant log preprocessing strategies, (2) dataset simplicity, and (3) the nature of binary classification in log anomaly detection. To assess the necessity of DL, we propose LightAD, an architecture that optimizes training time, inference time, and performance score. With automated hyper-parameter tuning, LightAD allows fair comparisons among log anomaly detection models, enabling engineers to evaluate the suitability of complex DL methods.

Our findings serve as a cautionary tale for the log anomaly detection community, highlighting the need to critically analyze datasets and research tasks before adopting DL approaches. Researchers proposing computationally expensive models should benchmark their work against lightweight algorithms to ensure a comprehensive evaluation.

Thu 18 Apr

Displayed time zone: Lisbon change

11:00 - 12:30
Language Models and Generated Code 2Demonstrations / Research Track at Maria Helena Vieira da Silva
Chair(s): Reyhaneh Jabbarvand University of Illinois at Urbana-Champaign
11:00
15m
Talk
Exploring the Potential of ChatGPT in Automated Code Refinement: An Empirical Study
Research Track
Qi Guo Tianjin University, China, Junming Cao Fudan University, Xiaofei Xie Singapore Management University, Shangqing Liu Nanyang Technological University, Xiaohong Li Tianjin University, Bihuan Chen Fudan University, Xin Peng Fudan University
11:15
15m
Talk
Deep Learning or Classical Machine Learning? An Empirical Study on Log-Based Anomaly Detection
Research Track
BoXi Yu The Chinese University of Hong Kong, Shenzhen, Jiayi Yao The Chinese University of Hong Kong, Shenzhen, Qiuai Fu Huawei Cloud Computing Technologies CO., LTD., Zhiqing Zhong Chinese University of Hong Kong, Shenzhen, Haotian Xie The Chinese University of Hong Kong, Shenzhen, Yaoliang Wu Huawei Cloud Computing Technologies Co., Ltd., Yuchi Ma Huawei Cloud Computing Technologies CO., LTD., Pinjia He Chinese University of Hong Kong, Shenzhen
11:30
15m
Talk
TRACED: Execution-aware Pre-training for Source Code
Research Track
Yangruibo Ding Columbia University, Benjamin Steenhoek Iowa State University, Kexin Pei The University of Chicago, Gail Kaiser Columbia University, Wei Le Iowa State University, Baishakhi Ray AWS AI Labs
11:45
15m
Talk
On Extracting Specialized Code Abilities from Large Language Models: A Feasibility Study
Research Track
Li Zongjie Hong Kong University of Science and Technology, Chaozheng Wang The Chinese University of Hong Kong, Pingchuan Ma HKUST, Chaowei Liu National University of Singapore, Shuai Wang The Hong Kong University of Science and Technology, Daoyuan Wu Nanyang Technological University, Cuiyun Gao Harbin Institute of Technology, Yang Liu Nanyang Technological University
12:00
15m
Talk
When Neural Code Completion Models Size up the Situation: Attaining Cheaper and Faster Completion through Dynamic Model Inference
Research Track
Zhensu Sun Singapore Management University, Xiaoning Du Monash University, Australia, Fu Song State Key Laboratory of Computer Science and Institute of Software, Chinese Academy of Sciences., Shangwen Wang National University of Defense Technology, Li Li Beihang University
Pre-print
12:15
7m
Talk
TestSpark: IntelliJ IDEA’s Ultimate Test Generation Companion
Demonstrations
Arkadii Sapozhnikov JetBrains Research, Mitchell Olsthoorn Delft University of Technology, Annibale Panichella Delft University of Technology, Vladimir Kovalenko JetBrains Research, Pouria Derakhshanfar JetBrains Research