Building robust prediction models for defective sensor data using Artificial Neural Networks

Predicting the health of components in complex dynamic systems such as an automobile poses numerous challenges. The primary aim of such predictive systems is to use the high-dimensional data acquired from different sensors and predict the state-of-health of a particular component, e.g., brake pad. The classical approach involves selecting a smaller set of relevant sensor signals using feature selection and using them to train a machine learning algorithm. However, this fails to address two prominent problems: (1) sensors are susceptible to failure when exposed to extreme conditions over a long periods of time; (2) sensors are electrical devices that can be affected by noise or electrical interference. Using the failed and noisy sensor signals as inputs largely reduce the prediction accuracy. To tackle this problem, it is advantageous to use the information from all sensor signals, so that the failure of one sensor can be compensated by another. In this work, we propose an Artificial Neural Network (ANN) based framework to exploit the information from a large number of signals. Secondly, our framework introduces a data augmentation approach to perform accurate predictions in spite of noisy signals. The plausibility of our framework is validated on real life industrial application from Robert Bosch GmbH.

[1]  Yoshua Bengio,et al.  An empirical analysis of dropout in piecewise linear networks , 2013, ICLR.

[2]  Andrian Marcus,et al.  Data Cleansing: Beyond Integrity Analysis , 2000, IQ.

[3]  Andrian Marcus,et al.  Data Cleansing: Beyond Integrity Analysis 1 , 2000 .

[4]  Xingquan Zhu,et al.  Class Noise vs. Attribute Noise: A Quantitative Study , 2003, Artificial Intelligence Review.

[5]  Peter S. Maybeck,et al.  Stochastic Models, Estimation And Control , 2012 .

[6]  Achilleas Zapranis,et al.  Stock performance modeling using neural networks: A comparative study with regression models , 1994, Neural Networks.

[7]  Brigitte Chebel-Morello,et al.  Accurate bearing remaining useful life prediction based on Weibull distribution and artificial neural network , 2015 .

[8]  Arvind Kumar Shekar,et al.  Including Multi-feature Interactions and Redundancy for Feature Ranking in Mixed Datasets , 2017, ECML/PKDD.

[9]  Alexander J. Smola,et al.  Support Vector Regression Machines , 1996, NIPS.

[10]  J V Tu,et al.  Advantages and disadvantages of using artificial neural networks versus logistic regression for predicting medical outcomes. , 1996, Journal of clinical epidemiology.

[11]  Sepp Hochreiter,et al.  Self-Normalizing Neural Networks , 2017, NIPS.

[12]  Nitish Srivastava,et al.  Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..

[13]  Leo Breiman,et al.  Random Forests , 2001, Machine Learning.

[14]  Ferat Sahin,et al.  A survey on feature selection methods , 2014, Comput. Electr. Eng..

[15]  D. Allred,et al.  Prognostic and predictive factors in breast cancer by immunohistochemical analysis. , 1998, Modern pathology : an official journal of the United States and Canadian Academy of Pathology, Inc.

[16]  Andrew Zisserman,et al.  Three things everyone should know to improve object retrieval , 2012, 2012 IEEE Conference on Computer Vision and Pattern Recognition.

[17]  Holger R. Maier,et al.  Neural networks for the prediction and forecasting of water resource variables: a review of modelling issues and applications , 2000, Environ. Model. Softw..

[18]  Arvind Kumar Shekar,et al.  Diverse Selection of Feature Subsets for Ensemble Regression , 2017, DaWaK.

[19]  Jocelyn Sietsma,et al.  Creating artificial neural networks that generalize , 1991, Neural Networks.

[20]  Khaled Elleithy,et al.  Innovations and Advances in Computer, Information, Systems Sciences, and Engineering , 2013 .

[21]  Miguel Lázaro Gredilla Sparse gaussian processes for large-scale machine learning , 2011 .

[22]  Bernard Widrow,et al.  Neural networks: applications in industry, business and science , 1994, CACM.

[23]  Veda C. Storey,et al.  A Framework for Analysis of Data Quality Research , 1995, IEEE Trans. Knowl. Data Eng..

[24]  Alberto Maria Segre,et al.  Programs for Machine Learning , 1994 .

[25]  Lluís A. Belanche Muñoz,et al.  Feature selection algorithms: a survey and experimental evaluation , 2002, 2002 IEEE International Conference on Data Mining, 2002. Proceedings..

[26]  Thomas Redman,et al.  Data quality for the information age , 1996 .

[27]  Luis Perez,et al.  The Effectiveness of Data Augmentation in Image Classification using Deep Learning , 2017, ArXiv.

[28]  Mieczysław Dziubiński,et al.  Electromagnetic interference in electrical systems of motor vehicles , 2016 .

[29]  Christoph Steup,et al.  Assessing neural networks for sensor fault detection , 2014, 2014 IEEE International Conference on Computational Intelligence and Virtual Environments for Measurement Systems and Applications (CIVEMSA).

[30]  William G. Baxt,et al.  Use of an Artificial Neural Network for Data Analysis in Clinical Decision-Making: The Diagnosis of Acute Coronary Occlusion , 1990, Neural Computation.

[31]  Pascal Reuss,et al.  Knowledge Engineering for Decision Support on Diagnosis and Maintenance in the Aircraft Domain , 2018, Synergies Between Knowledge Engineering and Software Engineering.

[32]  Yoshua Bengio,et al.  Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation , 2014, EMNLP.

[33]  Morteza Esfandyari,et al.  Stock Market Index Prediction Using Artificial Neural Network , 2016 .