Modular Neural Networks for Low-Power Image Classification on Embedded Devices

Embedded devices are generally small, battery-powered computers with limited hardware resources. It is difficult to run deep neural networks (DNNs) on these devices, because DNNs perform millions of operations and consume significant amounts of energy. Prior research has shown that a considerable number of a DNN’s memory accesses and computation are redundant when performing tasks like image classification. To reduce this redundancy and thereby reduce the energy consumption of DNNs, we introduce the Modular Neural Network Tree architecture. Instead of using one large DNN for the classifier, this architecture uses multiple smaller DNNs (called modules) to progressively classify images into groups of categories based on a novel visual similarity metric. Once a group of categories is selected by a module, another module then continues to distinguish among the similar categories within the selected group. This process is repeated over multiple modules until we are left with a single category. The computation needed to distinguish dissimilar groups is avoided, thus reducing redundant operations, memory accesses, and energy. Experimental results using several image datasets reveal the effectiveness of our proposed solution to reduce memory requirements by 50% to 99%, inference time by 55% to 95%, energy consumption by 52% to 94%, and the number of operations by 15% to 99% when compared with existing DNN architectures, running on two different embedded systems: Raspberry Pi 3 and Raspberry Pi Zero.

[1]  Rich Caruana,et al.  Do Deep Nets Really Need to be Deep? , 2013, NIPS.

[2]  Trevor Hastie,et al.  The Elements of Statistical Learning , 2001 .

[3]  Natalie D. Enright Jerger,et al.  Cnvlutin: Ineffectual-Neuron-Free Deep Neural Network Computing , 2016, 2016 ACM/IEEE 43rd Annual International Symposium on Computer Architecture (ISCA).

[4]  Hao Yu,et al.  Distributed Machine Learning on Smart-Gateway Network toward Real-Time Smart-Grid Energy Management with Behavior Cognition , 2018, ACM Trans. Design Autom. Electr. Syst..

[5]  Xuemin Chen,et al.  Internet of video things in 2030: A world with many cameras , 2017, 2017 IEEE International Symposium on Circuits and Systems (ISCAS).

[6]  Kilian Q. Weinberger,et al.  Densely Connected Convolutional Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[7]  Gregory Cohen,et al.  EMNIST: an extension of MNIST to handwritten letters , 2017, CVPR 2017.

[8]  Geoffrey E. Hinton,et al.  Distilling the Knowledge in a Neural Network , 2015, ArXiv.

[9]  Oliver Durr,et al.  Speaker identification and clustering using convolutional neural networks , 2016, 2016 IEEE 26th International Workshop on Machine Learning for Signal Processing (MLSP).

[10]  Tao Zhang,et al.  A Survey of Model Compression and Acceleration for Deep Neural Networks , 2017, ArXiv.

[11]  Forrest N. Iandola,et al.  SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <1MB model size , 2016, ArXiv.

[12]  Suresh Padmanabhan,et al.  Visual positioning system for automated indoor/outdoor navigation , 2017, TENCON 2017 - 2017 IEEE Region 10 Conference.

[13]  Abhishek Das,et al.  Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization , 2016, 2017 IEEE International Conference on Computer Vision (ICCV).

[14]  Dumitru Erhan,et al.  Deep Neural Networks for Object Detection , 2013, NIPS.

[15]  Vito Di Gesù,et al.  Distance-based functions for image comparison , 1999, Pattern Recognit. Lett..

[16]  Yiran Chen,et al.  Three years of low-power image recognition challenge: Introduction to special session , 2018, 2018 Design, Automation & Test in Europe Conference & Exhibition (DATE).

[17]  George A. Miller,et al.  WordNet: A Lexical Database for English , 1995, HLT.

[18]  F Matera Modular neural networks. , 1998, Substance use & misuse.

[19]  Ali Farhadi,et al.  Attribute Discovery via Predictable Discriminative Binary Codes , 2012, ECCV.

[20]  Daphna Weinshall,et al.  Exploiting Object Hierarchy: Combining Models from Different Category Levels , 2007, 2007 IEEE 11th International Conference on Computer Vision.

[21]  Gang Wang,et al.  Learning fine-grained features via a CNN Tree for Large-scale Classification , 2015, Neurocomputing.

[22]  Olivier Gibaru,et al.  CNN features are also great at unsupervised classification , 2017, ArXiv.

[23]  Li Lin,et al.  Joint Hierarchical Category Structure Learning and Large-Scale Image Classification , 2017, IEEE Transactions on Image Processing.

[24]  Sergey Ioffe,et al.  Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning , 2016, AAAI.

[25]  Antonio Torralba,et al.  Ieee Transactions on Pattern Analysis and Machine Intelligence 1 80 Million Tiny Images: a Large Dataset for Non-parametric Object and Scene Recognition , 2022 .

[26]  Kilian Q. Weinberger,et al.  CondenseNet: An Efficient DenseNet Using Learned Group Convolutions , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[27]  Xuan Zeng,et al.  Optimization and Quality Estimation of Circuit Design via Random Region Covering Method , 2017, ACM Trans. Design Autom. Electr. Syst..

[28]  Yung-Hsiang Lu Low-power image recognition , 2019 .

[29]  Emmanuel Dufourq,et al.  EDEN: Evolutionary deep networks for efficient machine learning , 2017, 2017 Pattern Recognition Association of South Africa and Robotics and Mechatronics (PRASA-RobMech).

[30]  Pietro Perona,et al.  Learning and using taxonomies for fast visual categorization , 2008, 2008 IEEE Conference on Computer Vision and Pattern Recognition.

[31]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[32]  Sheldon X.-D. Tan,et al.  Thermal-Sensor-Based Occupancy Detection for Smart Buildings Using Machine-Learning Methods , 2018, ACM Trans. Design Autom. Electr. Syst..

[33]  Ali Farhadi,et al.  XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks , 2016, ECCV.

[34]  Nir Friedman,et al.  Bayesian Network Classifiers , 1997, Machine Learning.

[35]  Parami Wijesinghe,et al.  FALCON: Feature Driven Selective Classification for Energy-Efficient Image Recognition , 2016, IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems.

[36]  Priyadarshini Panda,et al.  Tree-CNN: A hierarchical Deep Convolutional Neural Network for incremental learning , 2018, Neural Networks.

[37]  Alex Krizhevsky,et al.  Learning Multiple Layers of Features from Tiny Images , 2009 .

[38]  Li Fei-Fei,et al.  ImageNet: A large-scale hierarchical image database , 2009, CVPR.

[39]  Samuel Kaski,et al.  Dimensionality reduction by random mapping: fast similarity computation for clustering , 1998, 1998 IEEE International Joint Conference on Neural Networks Proceedings. IEEE World Congress on Computational Intelligence (Cat. No.98CH36227).

[40]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[41]  Yurong Chen,et al.  Dynamic Network Surgery for Efficient DNNs , 2016, NIPS.

[42]  Joan Bruna,et al.  Exploiting Linear Structure Within Convolutional Networks for Efficient Evaluation , 2014, NIPS.

[43]  Song Han,et al.  Deep Compression: Compressing Deep Neural Network with Pruning, Trained Quantization and Huffman Coding , 2015, ICLR.

[44]  Peter Kontschieder,et al.  Deep Neural Decision Forests , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).

[45]  Haim Levkowitz,et al.  GLHS: A Generalized Lightness, Hue, and Saturation Color Model , 1993, CVGIP Graph. Model. Image Process..

[46]  Peter E. Hart,et al.  Nearest neighbor pattern classification , 1967, IEEE Trans. Inf. Theory.

[47]  George K. Thiruvathukal,et al.  Observing Responses to the COVID-19 Pandemic using Worldwide Network Cameras , 2020, ArXiv.

[48]  Hassan Ghasemzadeh,et al.  Trading Off Power Consumption and Prediction Performance in Wearable Motion Sensors , 2018, ACM Trans. Design Autom. Electr. Syst..

[49]  Jianmin Wang,et al.  Deep Hashing Network for Efficient Similarity Retrieval , 2016, AAAI.

[50]  T. N. Vijaykumar,et al.  SparTen: A Sparse Tensor Accelerator for Convolutional Neural Networks , 2019, MICRO.

[51]  George K. Thiruvathukal,et al.  Low-Power Computer Vision: Status, Challenges, and Opportunities , 2019, IEEE Journal on Emerging and Selected Topics in Circuits and Systems.

[52]  R. D. Blanton,et al.  CompactNet: High Accuracy Deep Neural Network Optimized for On-Chip Implementation , 2018, 2018 IEEE International Conference on Big Data (Big Data).

[53]  Hanjiang Lai,et al.  Supervised Hashing for Image Retrieval via Image Representation Learning , 2014, AAAI.

[54]  Qiang Xu,et al.  Energy-Efficient and Quality-Assured Approximate Computing Framework Using a Co-Training Method , 2019, ACM Trans. Design Autom. Electr. Syst..

[55]  Andrew Zisserman,et al.  Speeding up Convolutional Neural Networks with Low Rank Expansions , 2014, BMVC.

[56]  Jiajun Li,et al.  SynergyFlow , 2018, ACM Trans. Design Autom. Electr. Syst..

[57]  J. Ross Quinlan,et al.  Induction of Decision Trees , 1986, Machine Learning.

[58]  Vincent Lepetit,et al.  Learning Separable Filters , 2013, 2013 IEEE Conference on Computer Vision and Pattern Recognition.

[59]  John Langford,et al.  Conditional Probability Tree Estimation Analysis and Algorithms , 2009, UAI.

[60]  Cordelia Schmid,et al.  Constructing Category Hierarchies for Visual Recognition , 2008, ECCV.

[61]  Anil K. Jain,et al.  Object detection using gabor filters , 1997, Pattern Recognit..

[62]  Paolo Napoletano,et al.  Benchmark Analysis of Representative Deep Neural Network Architectures , 2018, IEEE Access.

[63]  Silvio Savarese,et al.  Find the Best Path: An Efficient and Accurate Classifier for Image Hierarchies , 2013, 2013 IEEE International Conference on Computer Vision.

[64]  Robert LIN,et al.  NOTE ON FUZZY SETS , 2014 .

[65]  Lei Zhang,et al.  Bit-Scalable Deep Hashing With Regularized Similarity Learning for Image Retrieval and Person Re-Identification , 2015, IEEE Transactions on Image Processing.

[66]  Xiaodong Yang,et al.  Discriminative Hierarchical K-Means Tree for Large-Scale Image Classification , 2015, IEEE Transactions on Neural Networks and Learning Systems.

[67]  G. Griffin,et al.  Caltech-256 Object Category Dataset , 2007 .

[68]  Ali Farhadi,et al.  YOLO9000: Better, Faster, Stronger , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[69]  Heng Tao Shen,et al.  Unsupervised Deep Hashing with Similarity-Adaptive and Discrete Optimization , 2018, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[70]  Mark Sandler,et al.  MobileNetV2: Inverted Residuals and Linear Bottlenecks , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[71]  Hanan Samet,et al.  Pruning Filters for Efficient ConvNets , 2016, ICLR.

[72]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[73]  Nikos Komodakis,et al.  Wide Residual Networks , 2016, BMVC.

[74]  George K. Thiruvathukal,et al.  Camera Placement Meeting Restrictions of Computer Vision , 2020, 2020 IEEE International Conference on Image Processing (ICIP).

[75]  Andrew Y. Ng,et al.  Reading Digits in Natural Images with Unsupervised Feature Learning , 2011 .

[76]  William J. Dally,et al.  SCNN: An accelerator for compressed-sparse convolutional neural networks , 2017, 2017 ACM/IEEE 44th Annual International Symposium on Computer Architecture (ISCA).

[77]  Alexander C. Berg,et al.  Fast and Balanced: Efficient Label Tree Learning for Large Scale Object Recognition , 2011, NIPS.

[78]  Quoc V. Le,et al.  EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks , 2019, ICML.

[79]  N. Singpurwalla,et al.  Membership Functions and Probability Measures of Fuzzy Sets , 2004 .

[80]  Andrew Zisserman,et al.  Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.

[81]  Sinisa Todorovic,et al.  Monocular Depth Estimation Using Neural Regression Forest , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[82]  Frank Vahid,et al.  Switching Predictive Control Using Reconfigurable State-Based Model , 2019, ACM Trans. Design Autom. Electr. Syst..

[83]  Eugenio Culurciello,et al.  An Analysis of Deep Neural Network Models for Practical Applications , 2016, ArXiv.

[84]  David W. Conrath,et al.  Semantic Similarity Based on Corpus Statistics and Lexical Taxonomy , 1997, ROCLING/IJCLCLP.

[85]  George K. Thiruvathukal,et al.  A Survey of Methods for Low-Power Deep Learning and Computer Vision , 2020, 2020 IEEE 6th World Forum on Internet of Things (WF-IoT).

[86]  Shih-Fu Chang,et al.  An Exploration of Parameter Redundancy in Deep Networks with Circulant Projections , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).

[87]  Xin Li,et al.  Training Fixed-Point Classifiers for On-Chip Low-Power Implementation , 2017, ACM Trans. Design Autom. Electr. Syst..

[88]  Valentino Peluso,et al.  Scalable-Effort ConvNets for Multilevel Classification , 2018, 2018 IEEE/ACM International Conference on Computer-Aided Design (ICCAD).

[89]  Tao Mei,et al.  Automatic Video Genre Categorization using Hierarchical SVM , 2006, 2006 International Conference on Image Processing.

[90]  Shih-Fu Chang,et al.  Semi-supervised hashing for scalable image retrieval , 2010, 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition.