Fascinating Supervisory Signals and Where to Find Them: Deep Anomaly Detection with Scale Learning

Due to the unsupervised nature of anomaly detection, the key to fueling deep models is finding supervisory signals. Different from current reconstruction-guided generative models and transformation-based contrastive models, we devise novel data-driven supervision for tabular data by introducing a characteristic -- scale -- as data labels. By representing varied sub-vectors of data instances, we define scale as the relationship between the dimensionality of original sub-vectors and that of representations. Scales serve as labels attached to transformed representations, thus offering ample labeled data for neural network training. This paper further proposes a scale learning-based anomaly detection method. Supervised by the learning objective of scale distribution alignment, our approach learns the ranking of representations converted from varied subspaces of each data instance. Through this proxy task, our approach models inherent regularities and patterns within data, which well describes data"normality". Abnormal degrees of testing instances are obtained by measuring whether they fit these learned patterns. Extensive experiments show that our approach leads to significant improvement over state-of-the-art generative/contrastive anomaly detection methods.

[1]  Marius Kloft,et al.  E$^{3}$3Outlier: a Self-Supervised Framework for Unsupervised Deep Outlier Detection , 2022, IEEE Trans. Pattern Anal. Mach. Intell..

[2]  Yue Zhao,et al.  ADBench: Anomaly Detection Benchmark , 2022, NeurIPS.

[3]  Yijie Wang,et al.  Deep Isolation Forest for Anomaly Detection , 2022, IEEE Transactions on Knowledge and Data Engineering.

[4]  Maja R. Rudolph,et al.  Latent Outlier Exposure for Anomaly Detection with Contaminated Data , 2022, ICML.

[5]  Yijie Wang,et al.  Effective Anomaly Detection Based on Reinforcement Learning in Network Traffic Data , 2021, 2021 IEEE 27th International Conference on Parallel and Distributed Systems (ICPADS).

[6]  Jiayu Zhou,et al.  RCA: A Deep Collaborative Autoencoder Approach for Anomaly Detection , 2021, IJCAI.

[7]  Yi Tay,et al.  SCARF: Self-Supervised Contrastive Learning using Random Feature Corruption , 2021, ICLR.

[8]  Zheng Zhang,et al.  Anomaly detection using improved deep SVDD model with data structure preservation , 2021, Pattern Recognit. Lett..

[9]  Ning Liu,et al.  Beyond Outlier Detection: Outlier Interpretation by Attention-Guided Triplet Deviation Network , 2021, WWW.

[10]  Tomas Pfister,et al.  CutPaste: Self-Supervised Learning for Anomaly Detection and Localization , 2021, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[11]  Maja R. Rudolph,et al.  Neural Transformation Learning for Deep Anomaly Detection Beyond Images , 2021, ICML.

[12]  Prateek Mittal,et al.  SSD: A Unified Framework for Self-Supervised Outlier Detection , 2021, ICLR.

[13]  Thomas G. Dietterich,et al.  A Unifying Review of Deep and Shallow Anomaly Detection , 2020, Proceedings of the IEEE.

[14]  Jinwoo Shin,et al.  CSI: Novelty Detection via Contrastive Learning on Distributionally Shifted Instances , 2020, NeurIPS.

[15]  Marius Kloft,et al.  Explainable Deep One-Class Classification , 2020, ICLR.

[16]  Yedid Hoshen,et al.  Classification-Based Anomaly Detection for General Data , 2020, ICLR.

[17]  Chunhua Shen,et al.  Self-Trained Deep Ordinal Regression for End-to-End Video Anomaly Detection , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[18]  Harsha Vardhan Simhadri,et al.  DROCC: Deep Robust One-Class Classification , 2020, ICML.

[19]  Natalia Gimelshein,et al.  PyTorch: An Imperative Style, High-Performance Deep Learning Library , 2019, NeurIPS.

[20]  Yongjun Wang,et al.  MIX: A Joint Learning Framework for Detecting Both Clustered and Scattered Outliers in Mixed-Type Data , 2019, 2019 IEEE International Conference on Data Mining (ICDM).

[21]  Anton van den Hengel,et al.  Deep Anomaly Detection with Deviation Networks , 2019, KDD.

[22]  Yue Zhao,et al.  PyOD: A Python Toolbox for Scalable Outlier Detection , 2019, J. Mach. Learn. Res..

[23]  Meng Wang,et al.  Generative Adversarial Active Learning for Unsupervised Outlier Detection , 2018, IEEE Transactions on Knowledge and Data Engineering.

[24]  Thomas G. Dietterich,et al.  Deep Anomaly Detection with Outlier Exposure , 2018, ICLR.

[25]  Alexander Binder,et al.  Deep One-Class Classification , 2018, ICML.

[26]  Ran El-Yaniv,et al.  Deep Anomaly Detection Using Geometric Transformations , 2018, NeurIPS.

[27]  Randy C. Paffenroth,et al.  Anomaly Detection with Robust Deep Autoencoders , 2017, KDD.

[28]  Michael E. Houle,et al.  On the evaluation of unsupervised outlier detection: measures, datasets, and an empirical study , 2016, Data Mining and Knowledge Discovery.

[29]  Ole Winther,et al.  Autoencoding beyond pixels using a learned similarity metric , 2015, ICML.

[30]  Gang Hua,et al.  Learning Discriminative Reconstructions for Unsupervised Outlier Removal , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).

[31]  Charu C. Aggarwal,et al.  Outlier Analysis , 2013, Springer New York.

[32]  Zhi-Hua Zhou,et al.  Isolation Forest , 2008, 2008 Eighth IEEE International Conference on Data Mining.

[33]  David H. Wolpert,et al.  No free lunch theorems for optimization , 1997, IEEE Trans. Evol. Comput..

[34]  Kishan G. Mehrotra,et al.  An improved algorithm for neural network classification of imbalanced training sets , 1993, IEEE Trans. Neural Networks.

[35]  Lior Wolf,et al.  Anomaly Detection for Tabular Data with Internal Contrastive Learning , 2022, ICLR.

[36]  Mihaela van der Schaar,et al.  VIME: Extending the Success of Self- and Semi-supervised Learning to Tabular Domain , 2020, NeurIPS.

[37]  Charu C. Aggarwal,et al.  Outlier Detection with Autoencoder Ensembles , 2017, SDM.