Parameter-Efficient Log Anomaly Detection based on Pre-training model and LORA

Shiming He* (Corresponding Author), Ying Lei, Ying Zhang, Kun Xie, Pradip Kumar Sharma

*Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingPublished conference contribution

Abstract

Logs record both the normal and abnormal system operating status at any time, which are crucial data during system operation. Log anomaly detection can help with system debugging and analyzing root causes, such as system fault, shutdown fault, null-pointer exception, illegal-argument exception, and class cast exception. Deep learning is widely applied to log anomaly detection to enhance detection accuracy. However, the deep learning model requires a lot of label logs, which consume large amounts of labor and time. To tackle this label requirement problem, the pre-training model is introduced, for instance, the Bidirectional Encoder Representations from Transformers (BERT). However, the pre-training model brings new problems. The parameters of BERT needed to be fine-tuned are huge, resulting in a high training overhead. Besides, the direct word sequence input representation of BERT ignores the semantic information among logs. Therefore, we propose a parameter-efficient log anomaly detection scheme (LogBP-LORA) based on BERT and Low-Rank Adaptation (LORA). LORA is an effective parameter-tuning strategy. LogBP-LORA increases bypass weight matrices and only updates the bypass parameters instead of all the original parameters to reduce the training overhead. Additionally, LogBP-LORA exploits log event sequence representation to obtain more semantic information with a shorter sequence length. Extensive experiments carry on three public log datasets, BGL, Thunderbird, and HDFS, demonstrate LogBP-LORA can obtain favorable performance with lower resource consumption. When fewer label data is available, LogBP-LORA achieves about 10%-99% higher F1-score compared with Neurallog, Deeplog, MADDC, and Loganomaly. The training parameters of LogBP-LoRA are only 0.06% of the original parameters of BERT.

Original languageEnglish
Title of host publicationProceedings - 2023 IEEE 34th International Symposium on Software Reliability Engineering, ISSRE 2023
PublisherIEEE Computer Society
Pages207-217
Number of pages11
ISBN (Electronic)9798350315943
DOIs
Publication statusPublished - 2 Nov 2023
Event34th IEEE International Symposium on Software Reliability Engineering, ISSRE 2023 - Florence, Italy
Duration: 9 Oct 202312 Oct 2023

Publication series

NameProceedings - International Symposium on Software Reliability Engineering, ISSRE
ISSN (Print)1071-9458

Conference

Conference34th IEEE International Symposium on Software Reliability Engineering, ISSRE 2023
Country/TerritoryItaly
CityFlorence
Period9/10/2312/10/23

Keywords

  • BERT
  • Log anomaly detection
  • log event
  • log feature extraction
  • parameter-tuning strategy

Fingerprint

Dive into the research topics of 'Parameter-Efficient Log Anomaly Detection based on Pre-training model and LORA'. Together they form a unique fingerprint.

Cite this