A Linear NAS Service of ConvNets for Fast Deployment in the Edge of 5G Networks

The 5G network brings about significant convenience in deploying neural network models to edge devices. However, the flexibility challenges the current generation of neural network architectures that seldom involve the target platform as the bounds during the structure searching. The difficult deployment is rooted in the hidden resource provision of the target devices during the SW/HW joint tuning, leading to mismatching between network models and system configuration. This work proposes a scalable neural network search service in the 5G environment to support a continuous knob of the network scales, by which the channel groups can overlap with each other to share the features with continuous coverage. It is proved that the proposed dimension of network scaling provides good predictability of the model performance on specific platforms, which can be further utilized to simplify the regular network architectural search procedures. Then we design a SW/HW co-design workflow that involves both the cloud and edge to fully utilize computing resources on target platforms, meanwhile keeping the network size as small as possible to save the provision of resources. The experimental results show that, with our scalable search service, the key metrics of the network model enjoy a continuous, monotonic, and linear function to the proposed hyper-parameter. The deployment to the Raspberry Pi board shows that the proposed method accurately controls both precision and size of the models; meanwhile, the corresponding search workflow successfully finds the proper network scales with 65 percent reduction of the NAS routines.

[1]  Song Han,et al.  Deep Compression: Compressing Deep Neural Network with Pruning, Trained Quantization and Huffman Coding , 2015, ICLR.

[2]  Liang Lin,et al.  SNAS: Stochastic Neural Architecture Search , 2018, ICLR.

[3]  Xiangyu Zhang,et al.  ShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[4]  Gang Feng,et al.  On Robustness of Network Slicing for Next-Generation Mobile Networks , 2019, IEEE Transactions on Communications.

[5]  Bo Chen,et al.  MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications , 2017, ArXiv.

[6]  Vijay Vasudevan,et al.  Learning Transferable Architectures for Scalable Image Recognition , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[7]  Quoc V. Le,et al.  Efficient Neural Architecture Search via Parameter Sharing , 2018, ICML.

[8]  Bo Chen,et al.  MnasNet: Platform-Aware Neural Architecture Search for Mobile , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[9]  Song Han,et al.  ProxylessNAS: Direct Neural Architecture Search on Target Task and Hardware , 2018, ICLR.

[10]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[11]  Li Fei-Fei,et al.  ImageNet: A large-scale hierarchical image database , 2009, CVPR.

[12]  Yiming Yang,et al.  DARTS: Differentiable Architecture Search , 2018, ICLR.

[13]  Yuandong Tian,et al.  FBNet: Hardware-Aware Efficient ConvNet Design via Differentiable Neural Architecture Search , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[14]  Keke Gai,et al.  Resource Management in Sustainable Cyber-Physical Systems Using Heterogeneous Cloud Computing , 2018, IEEE Transactions on Sustainable Computing.

[15]  Bo Zhang,et al.  FairNAS: Rethinking Evaluation Fairness of Weight Sharing Neural Architecture Search , 2019, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).