Deep Neural Networks (DNNs) have enabled state-of-the-art accuracy on many challenging artificial intelligence tasks. While most of the computation currently resides in the cloud, it is desirable to embed DNN processing locally near the sensor due to privacy, security, and latency concerns or limitations in communication bandwidth. Accordingly, there has been increasing interest in the research community to design energy-efficient DNNs. However, estimating energy consumption from the DNN model is much more difficult than other metrics such as storage cost (model size) and throughput (number of operations). This is due to the fact that a significant portion of the energy is consumed by data movement, which is difficult to extract directly from the DNN model. This work proposes an energy estimation methodology that can estimate the energy consumption of a DNN based on its architecture, sparsity, and bitwidth. This methodology can be used to evaluate the various DNN architectures and energy-efficient techniques that are currently being proposed in the field and guide the design of energy-efficient DNNs. We have released an online version of the energy estimation tool at energyestimation.mit.edu. We believe that this method will play a critical role in bridging the gap between algorithm and hardware design and provide useful insights for the development of energy-efficient DNNs.
[1]
Dumitru Erhan,et al.
Going deeper with convolutions
,
2014,
2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[2]
Forrest N. Iandola,et al.
SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <1MB model size
,
2016,
ArXiv.
[3]
Vivienne Sze,et al.
Designing Energy-Efficient Convolutional Neural Networks Using Energy-Aware Pruning
,
2016,
2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[4]
Michael S. Bernstein,et al.
ImageNet Large Scale Visual Recognition Challenge
,
2014,
International Journal of Computer Vision.
[5]
Igor Carron,et al.
XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks
,
2016
.
[6]
Song Han,et al.
Deep Compression: Compressing Deep Neural Network with Pruning, Trained Quantization and Huffman Coding
,
2015,
ICLR.
[7]
Geoffrey E. Hinton,et al.
ImageNet classification with deep convolutional neural networks
,
2012,
Commun. ACM.