UniLog: Deploy One Model and Specialize it for All Log Analysis Tasks

Log analysis is vitally important for network reliability, and many log-based tasks are derived to analysis logs, such as anomaly detection, failure prediction, log compression, and log summarization. It is desired to have a unified log analysis framework to simultaneously run all these log analysis tasks on one model to achieve deployment convenience, superior task performance, and low maintenance cost. However, due to severe challenges about log data heterogeneity across devices, pioneer works design specialized algorithms for each task. In this work, we formulate the log analysis as a multi-task learning approach and propose to train a single model that can perform various log analysis tasks. We name this unified log analysis approach as UniLog. To effectively build an UniLog model, we propose a log data pretrained transformer to utilize the enormous unlabeled log data, and a corresponding multi-log-tasking finetune strategy for various log analysis tasks. Extensive experiments across seven datasets on four log analysis tasks demonstrate that UniLog achieves remarkable performance.

[1]  Frank Hutter,et al.  Fixing Weight Decay Regularization in Adam , 2017, ArXiv.

[2]  Yi Wang,et al.  Make A Long Image Short: Adaptive Token Length for Vision Transformers , 2021, ECML/PKDD.

[3]  Zibin Zheng,et al.  Tools and Benchmarks for Automated Log Parsing , 2018, 2019 IEEE/ACM 41st International Conference on Software Engineering: Software Engineering in Practice (ICSE-SEIP).

[4]  Rada Mihalcea,et al.  TextRank: Bringing Order into Text , 2004, EMNLP.

[5]  Yu Zhang,et al.  Log Clustering Based Problem Identification for Online Service Systems , 2016, 2016 IEEE/ACM 38th International Conference on Software Engineering Companion (ICSE-C).

[6]  Wei Xu,et al.  System Problem Detection by Mining Console Logs , 2010 .

[7]  Yann Dauphin,et al.  Language Modeling with Gated Convolutional Networks , 2016, ICML.

[8]  Shilin He,et al.  Loghub: A Large Collection of System Log Datasets towards Automated Log Analytics , 2020, ArXiv.

[9]  Quoc V. Le,et al.  Searching for Activation Functions , 2018, arXiv.

[10]  Ian H. Witten,et al.  Arithmetic coding for data compression , 1987, CACM.

[11]  Liwei Wang,et al.  On Layer Normalization in the Transformer Architecture , 2020, ICML.

[12]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[13]  Nitish Srivastava,et al.  Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..

[14]  Shilin He,et al.  Experience Report: System Log Analysis for Anomaly Detection , 2016, 2016 IEEE 27th International Symposium on Software Reliability Engineering (ISSRE).

[15]  Sebastian Ruder,et al.  Universal Language Model Fine-tuning for Text Classification , 2018, ACL.

[16]  Amos Storkey,et al.  Neural Architecture Search without Training , 2021, ICML.

[17]  Shenglin Zhang,et al.  Summarizing Unstructured Logs in Online Services , 2020, ArXiv.

[18]  知秀 柴田 5分で分かる!? 有名論文ナナメ読み:Jacob Devlin et al. : BERT : Pre-training of Deep Bidirectional Transformers for Language Understanding , 2020 .

[19]  Shenglin Zhang,et al.  LogParse: Making Log Parsing Adaptive through Word Classification , 2020, 2020 29th International Conference on Computer Communications and Networks (ICCCN).

[20]  Miroslaw Malek,et al.  Using Hidden Semi-Markov Models for Effective Online Failure Prediction , 2007, 2007 26th IEEE International Symposium on Reliable Distributed Systems (SRDS 2007).

[21]  Shenglin Zhang,et al.  LogTransfer: Cross-System Log Anomaly Detection for Software Systems with Transfer Learning , 2020, 2020 IEEE 31st International Symposium on Software Reliability Engineering (ISSRE).

[22]  Ilya Sutskever,et al.  Language Models are Unsupervised Multitask Learners , 2019 .

[23]  Di He,et al.  Do Transformers Really Perform Bad for Graph Representation? , 2021, ArXiv.

[24]  Ling Huang,et al.  Large-Scale System Problems Detection by Mining Console Logs , 2009 .

[25]  Glenn A. Fink,et al.  Predicting Computer System Failures Using Support Vector Machines , 2008, WASL.

[26]  Zibin Zheng,et al.  Logzip: Extracting Hidden Structures via Iterative Clustering for Log Compression , 2019, 2019 34th IEEE/ACM International Conference on Automated Software Engineering (ASE).

[27]  Jon Stearley,et al.  What Supercomputers Say: A Study of Five System Logs , 2007, 37th Annual IEEE/IFIP International Conference on Dependable Systems and Networks (DSN'07).

[28]  Yichen Zhu,et al.  Student Customized Knowledge Distillation: Bridging the Gap Between Student and Teacher , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).

[29]  Georg Heigold,et al.  An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale , 2021, ICLR.

[30]  Noah A. Smith,et al.  To Tune or Not to Tune? Adapting Pretrained Representations to Diverse Tasks , 2019, RepL4NLP@ACL.

[31]  Xintao Wu,et al.  LogBERT: Log Anomaly Detection via BERT , 2021, 2021 International Joint Conference on Neural Networks (IJCNN).

[32]  Qiang Fu,et al.  Mining Invariants from Console Logs for System Problem Detection , 2010, USENIX Annual Technical Conference.

[33]  Yang Liu,et al.  Fine-tune BERT for Extractive Summarization , 2019, ArXiv.

[34]  Feifei Li,et al.  DeepLog: Anomaly Detection and Diagnosis from System Logs through Deep Learning , 2017, CCS.

[35]  Colin Raffel,et al.  Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer , 2019, J. Mach. Learn. Res..

[36]  Shilin He,et al.  A Survey on Automated Log Analysis for Reliability Engineering , 2020, ACM Comput. Surv..

[37]  Mark Chen,et al.  Language Models are Few-Shot Learners , 2020, NeurIPS.

[38]  Shenglin Zhang,et al.  LogAnomaly: Unsupervised Detection of Sequential and Quantitative Anomalies in Unstructured Logs , 2019, IJCAI.

[39]  Lukasz Kaiser,et al.  Attention is All you Need , 2017, NIPS.

[40]  Geoffrey E. Hinton,et al.  A Simple Framework for Contrastive Learning of Visual Representations , 2020, ICML.