Evolutionary Neural Architecture Search Supporting Approximate Multipliers

There is a growing interest in automated neural architecture search (NAS) methods. They are employed to routinely deliver highquality neural network architectures for various challenging data sets and reduce the designer’s effort. The NAS methods utilizing multi-objective evolutionary algorithms are especially useful when the objective is not only to minimize the network error but also to minimize the number of parameters (weights) or power consumption of the inference phase. We propose a multi-objective NAS method based on Cartesian genetic programming for evolving convolutional neural networks (CNN). The method allows approximate operations to be used in CNNs to reduce power consumption of a target hardware implementation. During the NAS process, a suitable CNN architecture is evolved together with approximate multipliers to deliver the best trade-offs between the accuracy, network size and power consumption. The most suitable approximate multipliers are automatically selected from a library of approximate multipliers. Evolved CNNs are compared with common human-created CNNs of a similar complexity on the CIFAR-10 benchmark problem.

[1]  Frank Hutter,et al.  Neural Architecture Search: A Survey , 2018, J. Mach. Learn. Res..

[2]  Julian Francis Miller,et al.  Cartesian genetic programming , 2000, GECCO '10.

[3]  Yoshua Bengio,et al.  Random Search for Hyper-Parameter Optimization , 2012, J. Mach. Learn. Res..

[4]  Li Fei-Fei,et al.  Progressive Neural Architecture Search , 2017, ECCV.

[5]  Kalyanmoy Deb,et al.  NSGA-Net: neural architecture search using multi-objective genetic algorithm , 2018, GECCO.

[6]  Kaushik Roy,et al.  Invited — Cross-layer approximations for neuromorphic computing: From devices to circuits and systems , 2016, 2016 53nd ACM/EDAC/IEEE Design Automation Conference (DAC).

[7]  Xin Yao,et al.  Evolving artificial neural networks , 1999, Proc. IEEE.

[8]  Kaushik Roy,et al.  Energy-Efficient Neural Computing with Approximate Multipliers , 2018, ACM J. Emerg. Technol. Comput. Syst..

[9]  Taghi M. Khoshgoftaar,et al.  A survey on Image Data Augmentation for Deep Learning , 2019, Journal of Big Data.

[10]  Hokchhay Tann,et al.  Lightweight Deep Neural Network Accelerators Using Approximate SW/HW Techniques , 2018, Approximate Circuits.

[11]  Lukás Sekanina,et al.  EvoApproxSb: Library of approximate adders and multipliers for circuit design and benchmarking of approximation methods , 2017, Design, Automation & Test in Europe Conference & Exhibition (DATE), 2017.

[12]  Jingtong Hu,et al.  Standing on the Shoulders of Giants: Hardware and Neural Architecture Co-Search With Hot Start , 2020, IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems.

[13]  L. Sekanina,et al.  TFApprox: Towards a Fast Emulation of DNN Approximate Hardware Accelerators on GPU , 2020, 2020 Design, Automation & Test in Europe Conference & Exhibition (DATE).

[14]  Vivienne Sze,et al.  Efficient Processing of Deep Neural Networks: A Tutorial and Survey , 2017, Proceedings of the IEEE.

[15]  Wei Wei,et al.  2019 Formatting Instructions for Authors Using LaTeX , 2018 .

[16]  Risto Miikkulainen,et al.  Designing neural networks through neuroevolution , 2019, Nat. Mach. Intell..

[17]  Soheil Ghiasi,et al.  Ristretto: A Framework for Empirical Study of Resource-Efficient Inference in Convolutional Neural Networks , 2018, IEEE Transactions on Neural Networks and Learning Systems.

[18]  Yiming Yang,et al.  DARTS: Differentiable Architecture Search , 2018, ICLR.

[19]  Quoc V. Le,et al.  Large-Scale Evolution of Image Classifiers , 2017, ICML.

[20]  Risto Miikkulainen,et al.  Evolving Neural Networks through Augmenting Topologies , 2002, Evolutionary Computation.

[21]  Muhammad Shafique,et al.  An Updated Survey of Efficient Hardware Architectures for Accelerating Deep Convolutional Neural Networks , 2020, Future Internet.

[22]  Jian Sun,et al.  Identity Mappings in Deep Residual Networks , 2016, ECCV.

[23]  X. Yao Evolving Artificial Neural Networks , 1999 .

[24]  Quoc V. Le,et al.  Neural Architecture Search with Reinforcement Learning , 2016, ICLR.

[25]  Muhammad Shafique,et al.  ALWANN: Automatic Layer-Wise Approximation of Deep Neural Network Accelerators without Retraining , 2019, 2019 IEEE/ACM International Conference on Computer-Aided Design (ICCAD).

[26]  Martin Wistuba,et al.  A Survey on Neural Architecture Search , 2019, ArXiv.

[27]  Masanori Suganuma,et al.  A genetic programming approach to designing convolutional neural network architectures , 2017, GECCO.

[28]  Alan L. Yuille,et al.  Genetic CNN , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[29]  Kalyanmoy Deb,et al.  A fast and elitist multiobjective genetic algorithm: NSGA-II , 2002, IEEE Trans. Evol. Comput..