Incremental learning of neural network classifiers using reinforcement learning

With the availability of more data, classification is increasingly important. However, traditional classification algorithms do not scale well to large data sets and are often not suited when only limited samples of the dataset are available at any point in time. The latter arises, for example, in streaming data when the accumulation of data a priori is infeasible either due to limitations in memory or computation, or due to privacy and data ownership limitations. In these situations, traditional classification algorithms are difficult to apply since they are generally not incrementally trainable on changing data sets. To address this, this paper presents a novel approach that first uses Reinforcement Learning to learn a policy to incrementally build neural network classifiers for a broad distribution of problems and subsequently applies it to new data to learn a classifier for this specific problem. In both phases, learning operates on a sequence of small, randomly drawn subsets of the data, thus making it suitable for streaming data and for very large data sets where processing the entire set is not feasible. Experiments comparing this approach with kernel SVMs and large neural networks applied to the complete dataset show that this approach achieves comparable performance. Additional experiments were done to evaluate the performance of this approach for real world, streaming datasets and datasets with concept drift properties.

[1]  Leo Breiman,et al.  Bagging Predictors , 1996, Machine Learning.

[2]  Daniel Hernández-Lobato,et al.  An Analysis of Ensemble Pruning Techniques Based on Ordered Aggregation , 2009, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[3]  Yoav Freund,et al.  A decision-theoretic generalization of on-line learning and an application to boosting , 1995, EuroCOLT.

[4]  Eric Bauer,et al.  An Empirical Comparison of Voting Classification Algorithms: Bagging, Boosting, and Variants , 1999, Machine Learning.

[5]  Yoav Freund,et al.  Experiments with a New Boosting Algorithm , 1996, ICML.

[6]  Doina Precup,et al.  Between MDPs and Semi-MDPs: A Framework for Temporal Abstraction in Reinforcement Learning , 1999, Artif. Intell..

[7]  Pablo M. Granitto,et al.  Neural network ensembles: evaluation of aggregation algorithms , 2005, Artif. Intell..

[8]  William Nick Street,et al.  A streaming ensemble algorithm (SEA) for large-scale classification , 2001, KDD '01.

[9]  Jesús S. Aguilar-Ruiz,et al.  A similarity-based approach for data stream classification , 2014, Expert Syst. Appl..

[10]  Tamara G. Kolda,et al.  COMET: A Recipe for Learning and Using Large Ensembles on Massive Data , 2011, 2011 IEEE 11th International Conference on Data Mining.

[11]  Mahmood R. Azimi-Sadjadi,et al.  Recursive dynamic node creation in multilayer neural networks , 1993, IEEE Trans. Neural Networks.

[12]  Sattar Hashemi,et al.  Adapted One-versus-All Decision Trees for Data Stream Classification , 2009, IEEE Transactions on Knowledge and Data Engineering.

[13]  Salvatore J. Stolfo,et al.  On the Accuracy of Meta-learning for Scalable Data Mining , 2004, Journal of Intelligent Information Systems.

[14]  Stuart J. Russell,et al.  Online bagging and boosting , 2005, 2005 IEEE International Conference on Systems, Man and Cybernetics.

[15]  Wei Tang,et al.  Ensembling neural networks: Many could be better than all , 2002, Artif. Intell..

[16]  Richard S. Sutton,et al.  Generalization in Reinforcement Learning: Successful Examples Using Sparse Coarse Coding , 1995, NIPS.

[17]  Ryan M. Rifkin,et al.  In Defense of One-Vs-All Classification , 2004, J. Mach. Learn. Res..

[18]  William Nick Street,et al.  Ensemble Pruning Via Semi-definite Programming , 2006, J. Mach. Learn. Res..