Layers Sequence Optimizing for Deep Neural Networks using Multiples Objectives

Selecting the best architecture for a Deep Neural Network (DNN) is a non-trivial task since there is a massive amount of possible configurations (layers and parameters) and great difficulty in how to choose them. To make this task more independent of human interaction, this work addresses the DNN architecture selection problem as a multi-objective optimization task with different criteria in a combinatorial context. For this, we defined a new way to represent the architecture of DNN (layer sequence) as a solution in the optimization process. The proposed method attempts to find the best composition and sequence of layers for the DNN architecture satisfying two criteria: accuracy and $F_{1} {score}$. The method was evaluated for performance and compared to the exhaustive and random approaches and state-of-the-art DNN algorithms. The results obtained showed that the proposed method is capable of achieving results close to the optimum, and competitive when compared to those results reached by state of the art algorithms.

[1]  Steven R. Young,et al.  Optimizing deep learning hyper-parameters through an evolutionary algorithm , 2015, MLHPC@SC.

[2]  Lulu Zhang,et al.  Parameter Estimation for Coupled Hydromechanical Simulation of Dynamic Compaction Based on Pareto Multiobjective Optimization , 2015 .

[3]  Elliot Meyerson,et al.  Evolving Deep Neural Networks , 2017, Artificial Intelligence in the Age of Neural Networks and Brain Computing.

[4]  Yongjun Xu,et al.  Multi-objective Pruning for CNNs using Genetic Algorithm , 2019, ICANN.

[5]  Ming-Hsuan Yang,et al.  Multi-objective convolutional learning for face labeling , 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[6]  Hisao Ishibuchi,et al.  Evolutionary many-objective optimization: A short review , 2008, 2008 IEEE Congress on Evolutionary Computation (IEEE World Congress on Computational Intelligence).

[7]  Seetha Hari,et al.  Learning From Imbalanced Data , 2019, Advances in Computer and Electrical Engineering.

[8]  Kalyanmoy Deb,et al.  A fast and elitist multiobjective genetic algorithm: NSGA-II , 2002, IEEE Trans. Evol. Comput..

[9]  L. Deng,et al.  The MNIST Database of Handwritten Digit Images for Machine Learning Research [Best of the Web] , 2012, IEEE Signal Processing Magazine.

[10]  Nuno Lourenço,et al.  Evolving the Topology of Large Scale Deep Neural Networks , 2018, EuroGP.

[11]  Iddo Greental,et al.  Genetic algorithms for evolving deep neural networks , 2014, GECCO.

[12]  J. D. Schaffer,et al.  Combinations of genetic algorithms and neural networks: a survey of the state of the art , 1992, [Proceedings] COGANN-92: International Workshop on Combinations of Genetic Algorithms and Neural Networks.

[13]  Burt L. Monroe,et al.  Partial Justification of the Borda Count , 1998 .

[14]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[15]  José Ranilla,et al.  Particle swarm optimization for hyper-parameter selection in deep neural networks , 2017, GECCO.

[16]  H. Haenssle,et al.  Man against machine: diagnostic performance of a deep learning convolutional neural network for dermoscopic melanoma recognition in comparison to 58 dermatologists , 2018, Annals of oncology : official journal of the European Society for Medical Oncology.

[17]  Andrew Zisserman,et al.  Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.

[18]  Gaël Varoquaux,et al.  Scikit-learn: Machine Learning in Python , 2011, J. Mach. Learn. Res..

[19]  Frank Hutter,et al.  Neural Architecture Search: A Survey , 2018, J. Mach. Learn. Res..

[20]  Guigang Zhang,et al.  Deep Learning , 2016, Int. J. Semantic Comput..

[21]  Alex Krizhevsky,et al.  Learning Multiple Layers of Features from Tiny Images , 2009 .

[22]  Mohammad Teshnehlab,et al.  Evolutionary dynamic multi-objective optimization algorithm based on Borda count method , 2017, International Journal of Machine Learning and Cybernetics.

[23]  Jessica Barbosa Diniz,et al.  A Grammar-based Genetic Programming Approach to Optimize Convolutional Neural Network Architectures , 2018, Anais do XV Encontro Nacional de Inteligência Artificial e Computacional (ENIAC 2018).

[24]  Francisco Herrera,et al.  Advanced nonparametric tests for multiple comparisons in the design of experiments in computational intelligence and data mining: Experimental analysis of power , 2010, Inf. Sci..

[25]  Saeed Yahyanejad,et al.  Multi-Objective Convolutional Neural Networks for Robot Localisation and 3D Position Estimation in 2D Camera Images , 2018, 2018 15th International Conference on Ubiquitous Robots (UR).