Autoencoder-based One-class Classification Technique for Event Prediction

This paper proposes an autoencoder-based one-class classification technique to predict a specific event such as the occurrence of a fire in a specific building. Basically, a binary classification system that uses machine learning to identify fire-risk buildings requires 'positive' fire data and 'negative' non-fire data. However, the fire-risk building data that can be actually obtained have a single class data that includes only the data of the occurrence of the fire and does not include the data of the 'non-occurrence'. In this situation, PU (Positive-Unlabeled) learning which uses 'unlabeled' data can be an effective way of generating the fire prediction model. The autoencoder generates new features from the unlabeled data, with which a predictive model for predicting the fire-risk buildings is built through PU learning.

[1]  Malik Yousef,et al.  One-class document classification via Neural Networks , 2007, Neurocomputing.

[2]  Bing Liu,et al.  Learning with Positive and Unlabeled Examples Using Weighted Logistic Regression , 2003, ICML.

[3]  Han-joon Kim,et al.  Fire Prediction Technique using DNN-based Supervised Manifold Learning and Hierarchical Clustering Analysis , 2018 .

[4]  Duen Horng Chau,et al.  Firebird: Predicting Fire Risk and Prioritizing Fire Inspections in Atlanta , 2016, KDD.

[5]  Tie Luo,et al.  Distributed Anomaly Detection Using Autoencoder Neural Networks in WSN for IoT , 2018, 2018 IEEE International Conference on Communications (ICC).

[6]  Luca Benini,et al.  Anomaly Detection using Autoencoders in High Performance Computing Systems , 2018, DDC@AI*IA.

[7]  Ling Yu,et al.  Unsupervised Feature Selection Using RBF Autoencoder , 2019, ISNN.

[8]  Takehisa Yairi,et al.  Anomaly Detection Using Autoencoders with Nonlinear Dimensionality Reduction , 2014, MLSDA'14.

[9]  Gang Niu,et al.  Positive-Unlabeled Learning with Non-Negative Risk Estimator , 2017, NIPS.

[10]  Charles Elkan,et al.  Learning classifiers from only positive and unlabeled data , 2008, KDD.

[11]  Kwang-Kyu Seo,et al.  An application of one-class support vector machines in content-based image retrieval , 2007, Expert Syst. Appl..

[12]  Xiaoli Li,et al.  Learning from Positive and Unlabeled Examples with Different Data Distributions , 2005, ECML.

[13]  Wei Xiong,et al.  Stacked Convolutional Denoising Auto-Encoders for Feature Representation , 2017, IEEE Transactions on Cybernetics.

[14]  Geoffrey E. Hinton,et al.  Visualizing Data using t-SNE , 2008 .

[15]  Pascal Vincent,et al.  Stacked Denoising Autoencoders: Learning Useful Representations in a Deep Network with a Local Denoising Criterion , 2010, J. Mach. Learn. Res..

[16]  Bernhard Schölkopf,et al.  Estimating the Support of a High-Dimensional Distribution , 2001, Neural Computation.