A Hot/Cold Task Partition for Energy-Efficient Neural Network Deployment on Heterogeneous Edge Device

In recent years, neural network has made great achievements in the fields of image classification and object detection. At the same time, the rapid development of edge devices has also led many scholars to study the deployment of neural networks on resource-limited edge devices. However, the traditional neural network did not consider the specific application scenarios, which makes the data processing very inefficient. In this paper, we propose a method to deploy neural networks on heterogeneous edge devices. This method fully considers the specific applications scenarios of neural network, and proposes the idea of task partition. The whole dataset is divided into two classes: hot class and cold class, and then corresponding network models are trained respectively to handle the data. In order to speed up the execution efficiency, we propose a simple algorithm, which can balance the front-end and back-end load according to the processing capacity of edge devices. In addition, the time delay and energy consumption are considered in this paper. The experiments show that our method can save 36.4% energy consumption compared with the traditional method.

[1]  Li Fei-Fei,et al.  ImageNet: A large-scale hierarchical image database , 2009, CVPR.

[2]  David Hung-Chang Du,et al.  Distributed network computing over local ATM networks , 1994, Supercomputing '94.

[3]  Xu Chen,et al.  Edge Intelligence: On-Demand Deep Learning Model Co-Inference with Device-Edge Synergy , 2018, MECOMM@SIGCOMM.

[4]  Xiangyu Zhang,et al.  ShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[5]  Ming Yang,et al.  An Evaluation of the NVIDIA TX1 for Supporting Real-Time Computer-Vision Workloads , 2017, 2017 IEEE Real-Time and Embedded Technology and Applications Symposium (RTAS).

[6]  Beatriz Blanco-Filgueira,et al.  Deep Learning-Based Multiple Object Visual Tracking on Embedded System for IoT and Mobile Edge Computing Applications , 2018, IEEE Internet of Things Journal.

[7]  H. T. Kung,et al.  BranchyNet: Fast inference via early exiting from deep neural networks , 2016, 2016 23rd International Conference on Pattern Recognition (ICPR).

[8]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[9]  Andrew Zisserman,et al.  Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.

[10]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[11]  D. J. Finney Probit analysis; a statistical treatment of the sigmoid response curve. , 1947 .

[12]  C. E. SHANNON,et al.  A mathematical theory of communication , 1948, MOCO.

[13]  Alan Jay Smith,et al.  Cache Memories , 1982, CSUR.

[14]  Bo Chen,et al.  Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[15]  Bo Chen,et al.  MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications , 2017, ArXiv.