Hyperparameter Optimization in Convolutional Neural Network using Genetic Algorithms

Optimizing hyperparameters in Convolutional Neural Network (CNN) is a tedious problem for many researchers and practitioners. To get hyperparameters with better performance, experts are required to configure a set of hyperparameter choices manually. The best results of this manual configuration are thereafter modeled and implemented in CNN. However, different datasets require different model or combination of hyperparameters, which can be cumbersome and tedious. To address this, several works have been proposed such as grid search which is limited to low dimensional space, and tails which use random selection. Also, optimization methods such as evolutionary algorithms and Bayesian have been tested on MNIST datasets, which is less costly and require fewer hyperparameters than CIFAR-10 datasets. In this paper, the authors investigate the hyperparameter search methods on CIFAR-10 datasets. During the investigation with various optimization methods, performances in terms of accuracy are tested and recorded. Although there is no significant difference between propose approach and the state-of-the-art on CIFAR-10 datasets, however, the actual potency lies in the hybridization of genetic algorithms with local search method in optimizing both network structures and network training which is yet to be reported to the best of author knowledge.

[1]  Ian Dewancker,et al.  Bayesian Optimization for Machine Learning : A Practical Guidebook , 2016, ArXiv.

[2]  Michèle Sebag,et al.  Collaborative hyperparameter tuning , 2013, ICML.

[3]  David D. Cox,et al.  Making a Science of Model Search: Hyperparameter Optimization in Hundreds of Dimensions for Vision Architectures , 2013, ICML.

[4]  Prabhat,et al.  Scalable Bayesian Optimization Using Deep Neural Networks , 2015, ICML.

[5]  Aaron Klein,et al.  Fast Bayesian Optimization of Machine Learning Hyperparameters on Large Datasets , 2016, AISTATS.

[6]  Frank Hutter,et al.  Speeding Up Automatic Hyperparameter Optimization of Deep Neural Networks by Extrapolation of Learning Curves , 2015, IJCAI.

[7]  Himer Avila-George,et al.  Tuning the Parameters of a Convolutional Artificial Neural Network by Using Covering Arrays , 2016, Res. Comput. Sci..

[8]  Kevin Leyton-Brown,et al.  Sequential Model-Based Optimization for General Algorithm Configuration , 2011, LION.

[9]  Masaki Onishi,et al.  Effective hyperparameter optimization using Nelder-Mead method in deep learning , 2017, IPSJ Transactions on Computer Vision and Applications.

[10]  Lars Schmidt-Thieme,et al.  Hyperparameter Optimization Machines , 2016, 2016 IEEE International Conference on Data Science and Advanced Analytics (DSAA).

[11]  Dumitru Erhan,et al.  Going deeper with convolutions , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[12]  Quoc V. Le,et al.  On optimization methods for deep learning , 2011, ICML.

[13]  Katharina Eggensperger,et al.  Towards an Empirical Foundation for Assessing Bayesian Optimization of Hyperparameters , 2013 .

[14]  Ameet Talwalkar,et al.  Random Search and Reproducibility for Neural Architecture Search , 2019, UAI.

[15]  Lars Schmidt-Thieme,et al.  Learning hyperparameter optimization initializations , 2015, 2015 IEEE International Conference on Data Science and Advanced Analytics (DSAA).

[16]  Thomas Brox,et al.  Striving for Simplicity: The All Convolutional Net , 2014, ICLR.

[17]  Raúl Rojas,et al.  Neural Networks - A Systematic Introduction , 1996 .

[18]  Jasper Snoek,et al.  Practical Bayesian Optimization of Machine Learning Algorithms , 2012, NIPS.

[19]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[20]  Christophe G. Giraud-Carrier,et al.  Informing the Use of Hyperparameter Optimization Through Metalearning , 2017, 2017 IEEE International Conference on Data Mining (ICDM).

[21]  Alan L. Yuille,et al.  Genetic CNN , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[22]  Raymond Chiong,et al.  A Comparison between Genetic Algorithms and Evolutionary Programming based on Cutting Stock Problem , 2007, Eng. Lett..

[23]  Marc Parizeau,et al.  DEAP: evolutionary algorithms made easy , 2012, J. Mach. Learn. Res..

[24]  Tobias Senst,et al.  Hyper-parameter optimization for convolutional neural network committees based on evolutionary algorithms , 2017, 2017 IEEE International Conference on Image Processing (ICIP).

[25]  Qiang Chen,et al.  Network In Network , 2013, ICLR.

[26]  Abd. Rahman Ramli,et al.  Dynamic Bayesian networks and variable length genetic algorithm for designing cue-based model for dialogue act recognition , 2010, Comput. Speech Lang..

[27]  Yoshua Bengio,et al.  Maxout Networks , 2013, ICML.

[28]  Bart De Moor,et al.  Hyperparameter Search in Machine Learning , 2015, ArXiv.

[29]  Yoshua Bengio,et al.  Algorithms for Hyper-Parameter Optimization , 2011, NIPS.

[30]  Yoshua Bengio,et al.  Random Search for Hyper-Parameter Optimization , 2012, J. Mach. Learn. Res..

[31]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[32]  Zhuowen Tu,et al.  Deeply-Supervised Nets , 2014, AISTATS.

[33]  Yang Yuan,et al.  Hyperparameter Optimization: A Spectral Approach , 2017, ICLR.

[34]  Xin Yao,et al.  A new evolutionary system for evolving artificial neural networks , 1997, IEEE Trans. Neural Networks.