Joint Conference on Neural Networks , Orlando , Florida , USA , August 12-17 , 2007 PSMS for Neural Networks on the IJCNN 2007 Agnostic vs Prior Knowledge Challenge

Artificial neural networks have been proven to be effective learning algorithms since their introduction. These methods have been widely used in many domains, including scientific, medical, and commercial applications with great success. However, selecting the optimal combination of preprocessing methods and hyperparameters for a given data set is still a challenge. Recently a method for supervised learning model selection has been proposed: Particle Swarm Model Selection (PSMS). PSMS is a reliable method for the selection of optimal learning algorithms together with preprocessing methods, as well as for hyperparameter optimization. In this paper we applied PSMS for the selection of the (pseudo) optimal combination of preprocessing methods and hyperparameters for a fixed neural network on benchmark data sets from a challenging competition: the (IJCNN 2007) agnostic vs prior knowledge challenge. A forum for the evaluation of methods for model selection and data representation discovery. In this paper we further show that the use of PSMS is useful for model selection when we have no knowledge about the domain we are dealing with. With PSMS we obtained competitive models that are ranked high in the official results of the challenge.

[1]  Mauro Birattari,et al.  Swarm Intelligence , 2012, Lecture Notes in Computer Science.

[2]  Gavin C. Cawley,et al.  Agnostic Learning versus Prior Knowledge in the Design of Kernel Machines , 2007, 2007 International Joint Conference on Neural Networks.

[3]  Catherine Blake,et al.  UCI Repository of machine learning databases , 1998 .

[4]  Hugo Jair Escalante,et al.  Particle Swarm Model Selection , 2009, J. Mach. Learn. Res..

[5]  James Kennedy,et al.  Particle swarm optimization , 2002, Proceedings of ICNN'95 - International Conference on Neural Networks.

[6]  Russell C. Eberhart,et al.  Parameter Selection in Particle Swarm Optimization , 1998, Evolutionary Programming.

[7]  M. S. Voss,et al.  ARMA MODEL SELECTION USING PARTICLE SWARM OPTIMIZATION AND AIC CRITERIA , 2002 .

[8]  Isabelle Guyon,et al.  Agnostic Learning vs. Prior Knowledge Challenge , 2007, 2007 International Joint Conference on Neural Networks.

[9]  Nello Cristianini,et al.  Kernel Methods for Pattern Analysis , 2004 .

[10]  Roman W. Lutz,et al.  LogitBoost with Trees Applied to the WCCI 2006 Performance Prediction Challenge Datasets , 2006, The 2006 IEEE International Joint Conference on Neural Network Proceedings.

[11]  Juha Reunanen,et al.  Model Selection and Assessment Using Cross-indexing , 2007, 2007 International Joint Conference on Neural Networks.

[12]  M Reyes Sierra,et al.  Multi-Objective Particle Swarm Optimizers: A Survey of the State-of-the-Art , 2006 .

[13]  Hugo Jair Escalante,et al.  Particle Swarm Full Model Selection , 2008 .

[14]  I. Guyon,et al.  Performance Prediction Challenge , 2006, The 2006 IEEE International Joint Conference on Neural Network Proceedings.

[15]  J. Franklin,et al.  The elements of statistical learning: data mining, inference and prediction , 2005 .

[16]  O. Nelles Nonlinear System Identification: From Classical Approaches to Neural Networks and Fuzzy Models , 2000 .

[17]  Robert Tibshirani,et al.  The Elements of Statistical Learning: Data Mining, Inference, and Prediction, 2nd Edition , 2001, Springer Series in Statistics.