LLMParser: An Exploratory Study on Using Large Language Models for Log Parsing
Logs are important in modern software development with runtime information. Log parsing is the first step in many log-based analyses, that involve extracting structured information from unstructured log data. Traditional log parsers face challenges in accurately parsing logs due to the diversity of log formats, which directly impacts the performance of downstream log-analysis tasks. In this paper, we explore the potential of using Large Language Models (LLMs) for log parsing and propose LLMParser, an LLM-based log parser based on generative LLMs and few-shot tuning. We leverage four LLMs, Flan-T5-small, Flan-T5-base, LLaMA-7B, and ChatGLM-6B in LLMParsers. Our evaluation of 16 open-source systems shows that LLMParser achieves statistically significantly higher parsing accuracy than state-of-the-art parsers (a 96% average parsing accuracy). We further conduct a comprehensive empirical analysis on the effect of training size, model size, and pre-training LLM on log parsing accuracy. We find that smaller LLMs may be more effective than more complex LLMs, where Flan-T5-base achieves comparable results as LLaMA-7B with a shorter inference time. We also find that using LLMs pre-trained using logs from other systems does not always improve parsing accuracy. Using pre-trained Flan-T5-base shows an improvement in accuracy, but a decrease in pre-trained LLaMA (decrease by almost 55% in group accuracy). In short, our study provides empirical evidence for using LLMs for log parsing and highlights the limitations and future research direction of LLM-based log parsers.
Fri 19 AprDisplayed time zone: Lisbon change
11:00 - 12:30 | LLM, NN and other AI technologies 5Software Engineering Education and Training / Software Engineering in Practice / Research Track at Grande Auditório Chair(s): Baishakhi Ray AWS AI Labs | ||
11:00 15mTalk | Enhancing Exploratory Testing by Large Language Model and Knowledge Graph Research Track Yanqi Su Australian National University, Dianshu Liao Australian National University, Zhenchang Xing CSIRO's Data61, Qing Huang School of Computer Information Engineering, Jiangxi Normal University, Mulong Xie CSIRO's Data61, Qinghua Lu Data61, CSIRO, Xiwei (Sherry) Xu Data61, CSIRO | ||
11:15 15mTalk | LLMParser: An Exploratory Study on Using Large Language Models for Log Parsing Research Track Zeyang Ma Concordia University, An Ran Chen University of Alberta, Dong Jae Kim Concordia University, Tse-Hsun (Peter) Chen Concordia University, Shaowei Wang Department of Computer Science, University of Manitoba, Canada | ||
11:30 15mTalk | Enhancing Text-to-SQL Translation for Financial System Design Software Engineering in Practice Yewei Song University of Luxembourg, Saad Ezzini Lancaster University, Xunzhu Tang University of Luxembourg, Cedric Lothritz University of Luxembourg, Jacques Klein University of Luxembourg, Tegawendé F. Bissyandé University of Luxembourg, Andrey Boytsov Banque BGL BNP Paribas, Ulrick Ble Banque BGL BNP Paribas, Anne Goujon Banque BGL BNP Paribas | ||
11:45 15mTalk | Towards Building AI-CPS with NVIDIA Isaac Sim: An Industrial Benchmark and Case Study for Robotics Manipulation Software Engineering in Practice Zhehua Zhou University of Alberta, Jiayang Song University of Alberta, Xuan Xie University of Alberta, Zhan Shu University of Alberta, Lei Ma The University of Tokyo & University of Alberta, Dikai Liu NVIDIA AI Tech Centre, Jianxiong Yin NVIDIA AI Tech Centre, Simon See NVIDIA AI Tech Centre Pre-print | ||
12:00 15mTalk | Let's Ask AI About Their Programs: Exploring ChatGPT's Answers To Program Comprehension Questions Software Engineering Education and Training Pre-print Media Attached File Attached | ||
12:15 15mTalk | Experience Report: Identifying common misconceptions and errors of novice programmers with ChatGPT Software Engineering Education and Training Hua Leong Fwa Singapore Management University Media Attached |