MoNA: Mobile Neural Architecture with Reconfigurable Parallel Dimensions

Emerging mobile deep neural networks (DNNs) are prevalent in resource-limited devices because of their shrunken parameter size and reduced computation. However, when deploying mobile DNNs on conventional DNN accelerators, there will be a performance loss due to the reduction of processing element (PE) utilization. Conventional accelerators are generally designed base on a fixed pattern of data reuse, but the data reuse opportunities of different layers in mobile DNNs are diverse. To process mobile DNNs with high performance, we propose an architecture called MoNA (Mobile Neural Architecture), which includes a flexible dataflow and a reconfigurable computing core. The dataflow efficiently supports reconfigurable processing parallelism to maximize the PE utilization of mobile DNNs. The computing core is a 3D PE array, which supports the dataflow and avoids high bandwidth requirement.

[1]  Forrest N. Iandola,et al.  SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <1MB model size , 2016, ArXiv.

[2]  Xiangyu Zhang,et al.  ShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[3]  Leibo Liu,et al.  Deep Convolutional Neural Network Architecture With Reconfigurable Computation Patterns , 2017, IEEE Transactions on Very Large Scale Integration (VLSI) Systems.

[4]  Dumitru Erhan,et al.  Going deeper with convolutions , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[5]  Bo Chen,et al.  MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications , 2017, ArXiv.

[6]  David A. Patterson,et al.  In-datacenter performance analysis of a tensor processing unit , 2017, 2017 ACM/IEEE 44th Annual International Symposium on Computer Architecture (ISCA).

[7]  Mark Sandler,et al.  MobileNetV2: Inverted Residuals and Linear Bottlenecks , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[8]  Andrew Zisserman,et al.  Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.

[9]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[10]  Jia Wang,et al.  DaDianNao: A Machine-Learning Supercomputer , 2014, 2014 47th Annual IEEE/ACM International Symposium on Microarchitecture.