DANoC: An Efficient Algorithm and Hardware Codesign of Deep Neural Networks on Chip

Deep neural networks (NNs) are the state-of-the-art models for understanding the content of images and videos. However, implementing deep NNs in embedded systems is a challenging task, e.g., a typical deep belief network could exhaust gigabytes of memory and result in bandwidth and computational bottlenecks. To address this challenge, this paper presents an algorithm and hardware codesign for efficient deep neural computation. A hardware-oriented deep learning algorithm, named the deep adaptive network, is proposed to explore the sparsity of neural connections. By adaptively removing the majority of neural connections and robustly representing the reserved connections using binary integers, the proposed algorithm could save up to 99.9% memory utility and computational resources without undermining classification accuracy. An efficient sparse-mapping-memory-based hardware architecture is proposed to fully take advantage of the algorithmic optimization. Different from traditional Von Neumann architecture, the deep-adaptive network on chip (DANoC) brings communication and computation in close proximity to avoid power-hungry parameter transfers between on-board memory and on-chip computational units. Experiments over different image classification benchmarks show that the DANoC system achieves competitively high accuracy and efficiency comparing with the state-of-the-art approaches.

[1]  Song Han,et al.  Learning both Weights and Connections for Efficient Neural Network , 2015, NIPS.

[2]  Eddy Mayoraz,et al.  A constructive training algorithm for feedforward neural networks with ternary weights , 1994, ESANN.

[3]  Geoffrey E. Hinton A Practical Guide to Training Restricted Boltzmann Machines , 2012, Neural Networks: Tricks of the Trade.

[4]  Yoshua Bengio,et al.  Gradient-based learning applied to document recognition , 1998, Proc. IEEE.

[5]  Farhad Samadzadegan,et al.  Classification of Hyperspectral Data Using an AdaBoostSVM Technique Applied on Band Clusters , 2014, IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing.

[6]  Thomas S. Huang,et al.  Semisupervised Hyperspectral Classification Using Task-Driven Dictionary Learning With Laplacian Regularization , 2015, IEEE Transactions on Geoscience and Remote Sensing.

[7]  Igor Carron,et al.  XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks , 2016 .

[8]  Ian H. Witten,et al.  The WEKA data mining software: an update , 2009, SKDD.

[9]  David Saad,et al.  Training a network with ternary weights using the CHIR algorithm , 1993, IEEE Trans. Neural Networks.

[10]  Byungik Ahn,et al.  Computation of deep belief networks using special-purpose hardware architecture , 2014, 2014 International Joint Conference on Neural Networks (IJCNN).

[11]  Jason Cong,et al.  Optimizing FPGA-based Accelerator Design for Deep Convolutional Neural Networks , 2015, FPGA.

[12]  E. Culurciello,et al.  NeuFlow: Dataflow vision processing system-on-a-chip , 2012, 2012 IEEE 55th International Midwest Symposium on Circuits and Systems (MWSCAS).

[13]  R. Tibshirani,et al.  A note on the group lasso and a sparse group lasso , 2010, 1001.0736.

[14]  Song Han,et al.  Deep Compression: Compressing Deep Neural Network with Pruning, Trained Quantization and Huffman Coding , 2015, ICLR.

[15]  M. Siegel,et al.  Hyperspectral classification via deep networks and superpixel segmentation , 2015 .

[16]  Luca Benini,et al.  Origami: A 803-GOp/s/W Convolutional Network Accelerator , 2015, IEEE Transactions on Circuits and Systems for Video Technology.

[17]  M. Yuan,et al.  Model selection and estimation in regression with grouped variables , 2006 .

[18]  Honglak Lee,et al.  Sparse deep belief net model for visual area V2 , 2007, NIPS.

[19]  Antonio J. Plaza,et al.  A Subspace-Based Multinomial Logistic Regression for Hyperspectral Image Classification , 2014, IEEE Geoscience and Remote Sensing Letters.

[20]  Marc'Aurelio Ranzato,et al.  Sparse Feature Learning for Deep Belief Networks , 2007, NIPS.

[21]  James E. Fowler,et al.  Hyperspectral Image Classification Using Gaussian Mixture Models and Markov Random Fields , 2014, IEEE Geoscience and Remote Sensing Letters.

[22]  Qi Wang,et al.  Hyperspectral Band Selection by Multitask Sparsity Pursuit , 2015, IEEE Transactions on Geoscience and Remote Sensing.

[23]  James E. Fowler,et al.  Locality-Preserving Dimensionality Reduction and Classification for Hyperspectral Image Analysis , 2012, IEEE Transactions on Geoscience and Remote Sensing.

[24]  Tianshi Chen,et al.  ShiDianNao: Shifting vision processing closer to the sensor , 2015, 2015 ACM/IEEE 42nd Annual International Symposium on Computer Architecture (ISCA).

[25]  Berin Martini,et al.  A 240 G-ops/s Mobile Coprocessor for Deep Neural Networks , 2014, 2014 IEEE Conference on Computer Vision and Pattern Recognition Workshops.

[26]  S. Himavathi,et al.  Feedforward Neural Network Implementation in FPGA Using Layer Multiplexing for Effective Resource Utilization , 2007, IEEE Transactions on Neural Networks.

[27]  Xing Zhao,et al.  Spectral–Spatial Classification of Hyperspectral Data Based on Deep Belief Network , 2015, IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing.

[28]  Ran El-Yaniv,et al.  Binarized Neural Networks , 2016, ArXiv.

[29]  Shutao Li,et al.  Spectral–Spatial Hyperspectral Image Classification Based on KNN , 2016 .

[30]  Pritish Narayanan,et al.  Deep Learning with Limited Numerical Precision , 2015, ICML.

[31]  Sorin Draghici,et al.  On the capabilities of neural networks using limited precision weights , 2002, Neural Networks.

[32]  Bor-Chen Kuo,et al.  A Kernel-Based Feature Selection Method for SVM With RBF Kernel for Hyperspectral Image Classification , 2014, IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing.

[33]  Tara N. Sainath,et al.  Deep Neural Networks for Acoustic Modeling in Speech Recognition , 2012 .

[34]  Geoffrey E. Hinton,et al.  Reducing the Dimensionality of Data with Neural Networks , 2006, Science.

[35]  Shihong Du,et al.  Spectral–Spatial Feature Extraction for Hyperspectral Image Classification: A Dimension Reduction and Deep Learning Approach , 2016, IEEE Transactions on Geoscience and Remote Sensing.

[36]  Oriol Vinyals,et al.  Comparing multilayer perceptron to Deep Belief Network Tandem features for robust ASR , 2011, 2011 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[37]  Jenq-Neng Hwang,et al.  Finite Precision Error Analysis of Neural Network Hardware Implementations , 1993, IEEE Trans. Computers.

[38]  Ning Li,et al.  A multistage dataflow implementation of a Deep Convolutional Neural Network based on FPGA for high-speed object recognition , 2016, 2016 IEEE Southwest Symposium on Image Analysis and Interpretation (SSIAI).

[39]  Gang Wang,et al.  Deep Learning-Based Classification of Hyperspectral Data , 2014, IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing.

[40]  Yiran Chen,et al.  Learning Structured Sparsity in Deep Neural Networks , 2016, NIPS.

[41]  Mathieu Salzmann,et al.  Learning the Number of Neurons in Deep Networks , 2016, NIPS.

[42]  Ammar Belatreche,et al.  Challenges for large-scale implementations of spiking neural networks on FPGAs , 2007, Neurocomputing.

[43]  Noel Lopes,et al.  An Evaluation of Multiple Feed-Forward Networks on GPUs , 2011, Int. J. Neural Syst..

[44]  Andreas G. Andreou,et al.  FPGA implementation of a Deep Belief Network architecture for character recognition using stochastic computation , 2015, 2015 49th Annual Conference on Information Sciences and Systems (CISS).

[45]  Yixin Chen,et al.  Compressing Neural Networks with the Hashing Trick , 2015, ICML.

[46]  Shanjun Mao,et al.  Spectral–spatial classification of hyperspectral images using deep convolutional neural networks , 2015 .

[47]  Yike Guo,et al.  DropNeuron: Simplifying the Structure of Deep Neural Networks , 2016, ArXiv.

[48]  Chun-Xia Zhang,et al.  A sparse-response deep belief network based on rate distortion theory , 2014, Pattern Recognit..