Incremental model selection and ensemble prediction under virtual concept drifting environments

Model selection for machine learning systems is one of the most important issues to be addressed for obtaining greater generalization capabilities. This paper proposes a strategy to achieve model selection incrementally under virtual concept drifting environments, where the distribution of learning samples varies over time. To carry out incremental model selection, the system generally uses all the learning samples that have been observed until now. Under virtual concept drifting environments, however, the distribution of the observed samples is considerably different from the distribution of cumulative dataset so that model selection is usually unsuccessful. To overcome this problem, the author had earlier proposed the weighted objective function and model-selection criterion based on the predictive input density of the learning samples. Although the previous method described in the author’s previous study shows good performances to some datasets, it occasionally fails to yield appropriate learning results because of the failure in the prediction of the actual input density. To reduce the adverse effect, the method proposed in this paper improves on the previously described method to yield the desired outputs using an ensemble of the constructed radial basis function neural networks (RBFNNs).

[1]  Koichiro Yamauchi Incremental learning and model selection under virtual concept drifting environments , 2010, The 2010 International Joint Conference on Neural Networks (IJCNN).

[2]  Shigeo Abe,et al.  Incremental learning of feature space and classifier for face recognition , 2005, Neural Networks.

[3]  Koichiro Yamauchi Optimal incremental learning under covariate shift , 2009, Memetic Comput..

[4]  Narasimhan Sundararajan,et al.  A generalized growing and pruning RBF (GGAP-RBF) neural network for function approximation , 2005, IEEE Transactions on Neural Networks.

[5]  Bernard Ans,et al.  Neural networks with a self-refreshing memory: Knowledge transfer in sequential learning tasks without catastrophic forgetting , 2000, Connect. Sci..

[6]  Ezequiel López-Rubio,et al.  Multivariate Student-t self-organizing maps , 2009, Neural Networks.

[7]  Shin Ishii,et al.  On-line EM Algorithm for the Normalized Gaussian Network , 2000, Neural Computation.

[8]  H. Akaike A new look at the statistical model identification , 1974 .

[9]  James C. Bezdek,et al.  A Convergence Theorem for the Fuzzy ISODATA Clustering Algorithms , 1980, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[10]  Koichiro Yamauchi,et al.  Incremental Leaning and Model Selection for Radial Basis Function Network through Sleep , 2007, IEICE Trans. Inf. Syst..

[11]  Seiichi Ozawa,et al.  An Incremental Learning Algorithm for Resource Allocating Networks Based on Local Linear Regression , 2009, ICONIP.

[12]  H. Shimodaira,et al.  Improving predictive inference under covariate shift by weighting the log-likelihood function , 2000 .

[13]  Amaury Lendasse,et al.  Evolving fuzzy optimally pruned extreme learning machine for regression problems , 2010, Evol. Syst..

[14]  John C. Platt A Resource-Allocating Network for Function Interpolation , 1991, Neural Computation.

[15]  Motoaki Kawanabe,et al.  Direct Importance Estimation with Model Selection and Its Application to Covariate Shift Adaptation , 2007, NIPS.

[16]  Robert K. L. Gay,et al.  Error Minimized Extreme Learning Machine With Growth of Hidden Nodes and Incremental Learning , 2009, IEEE Transactions on Neural Networks.

[17]  Christopher G. Atkeson,et al.  Constructive Incremental Learning from Only Local Information , 1998, Neural Computation.

[18]  Koichiro Yamauchi Incremental learning and model selection under virtual concept drifting environments , 2010, Neuroscience Research.

[19]  Naohiro Ishii,et al.  Incremental learning methods with retrieving of interfered patterns , 1999, IEEE Trans. Neural Networks.

[20]  John Moody,et al.  Fast Learning in Networks of Locally-Tuned Processing Units , 1989, Neural Computation.

[21]  Robert M. French,et al.  Pseudo-recurrent Connectionist Networks: An Approach to the 'Sensitivity-Stability' Dilemma , 1997, Connect. Sci..

[22]  D. Rubin,et al.  Maximum likelihood from incomplete data via the EM - algorithm plus discussions on the paper , 1977 .

[23]  Yukinori Kakazu,et al.  Study on Optimization of Grinding Conditions using Neural Networks. A Method of Additional Learning. , 1992 .