RL-PDNN: Reinforcement Learning for Privacy-Aware Distributed Neural Networks in IoT Systems

Due to their high computational and memory demand, deep learning applications are mainly restricted to high-performance units, e.g., cloud and edge servers. Particularly, in Internet of Things (IoT) systems, the data acquired by pervasive devices is sent to the computing servers for classification. However, this approach might not be always possible because of the limited bandwidth and the privacy issues. Furthermore, it presents uncertainty in terms of latency because of the unstable remote connectivity. To support resource and delay requirements of such paradigm, joint and real-time deep co-inference framework with IoT synergy was introduced. However, scheduling the distributed, dynamic and real-time Deep Neural Network (DNN) inference requests among resource-constrained devices has not been well explored in the literature. Additionally, the distribution of DNN has drawn the attention to the privacy protection of sensitive data. In this context, various threats have been presented, including white-box attacks, where malicious devices can accurately recover received inputs if the DNN model is fully exposed to participants. In this paper, we introduce a methodology aiming at distributing the DNN tasks onto the resource-constrained devices of the IoT system, while avoiding to reveal the model to participants. We formulate such an approach as an optimization problem, where we establish a trade-off between the latency of co-inference, the privacy of the data, and the limited resources of devices. Next, due to the NP-hardness of the problem, we shape our approach as a reinforcement learning design adequate for real-time applications and highly dynamic systems, namely RL-PDNN. Our system proved its ability to outperform existing static approaches and achieve close results compared to the optimal solution.

[1]  H. T. Kung,et al.  Distributed Deep Neural Networks Over the Cloud, the Edge and End Devices , 2017, 2017 IEEE 37th International Conference on Distributed Computing Systems (ICDCS).

[2]  Cesare Alippi,et al.  Distributed Deep Convolutional Neural Networks for the Internet-of-Things , 2019, IEEE Transactions on Computers.

[3]  Christopher D. Manning,et al.  Effective Approaches to Attention-based Neural Machine Translation , 2015, EMNLP.

[4]  Michael S. Ryoo,et al.  Toward Collaborative Inferencing of Deep Neural Networks on Internet-of-Things Devices , 2020, IEEE Internet of Things Journal.

[5]  Trevor N. Mudge,et al.  Neurosurgeon: Collaborative Intelligence Between the Cloud and Mobile Edge , 2017, ASPLOS.

[6]  Shane Legg,et al.  Human-level control through deep reinforcement learning , 2015, Nature.

[7]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[8]  Xiaofei Wang,et al.  Convergence of Edge Computing and Deep Learning: A Comprehensive Survey , 2019, IEEE Communications Surveys & Tutorials.

[9]  Moustafa Alzantot,et al.  RSTensorFlow: GPU Enabled TensorFlow for Deep Learning on Commodity Android Devices , 2017, EMDL '17.

[10]  Ruby B. Lee,et al.  Model inversion attacks against collaborative inference , 2019, ACSAC.

[11]  Nicu Sebe,et al.  Binary Neural Networks: A Survey , 2020, Pattern Recognit..

[12]  Nikita Borisov,et al.  Property Inference Attacks on Fully Connected Neural Networks using Permutation Invariant Representations , 2018, CCS.

[13]  Xu Chen,et al.  Edge Intelligence: Paving the Last Mile of Artificial Intelligence With Edge Computing , 2019, Proceedings of the IEEE.

[14]  Andreas Gerstlauer,et al.  DeepThings: Distributed Adaptive Deep Learning Inference on Resource-Constrained IoT Edge Clusters , 2018, IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems.

[15]  Pritish Narayanan,et al.  Deep Learning with Limited Numerical Precision , 2015, ICML.

[16]  Massoud Pedram,et al.  JointDNN: An Efficient Training and Inference Engine for Intelligent Mobile Cloud Computing Services , 2018, IEEE Transactions on Mobile Computing.

[17]  Lingjuan Lyu,et al.  FORESEEN: Towards Differentially Private Deep Inference for Intelligent Internet of Things , 2020, IEEE Journal on Selected Areas in Communications.

[18]  Andrew Zisserman,et al.  Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.

[19]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[20]  Nicholas D. Lane,et al.  DeepX: A Software Accelerator for Low-Power Deep Learning Inference on Mobile Devices , 2016, 2016 15th ACM/IEEE International Conference on Information Processing in Sensor Networks (IPSN).

[21]  Zongpu Zhang,et al.  Towards Ubiquitous Intelligent Computing: Heterogeneous Distributed Deep Neural Networks , 2018, IEEE Transactions on Big Data.

[22]  Hyeong-Ju Kang,et al.  Accelerator-Aware Pruning for Convolutional Neural Networks , 2018, IEEE Transactions on Circuits and Systems for Video Technology.

[23]  Cesare Alippi,et al.  Moving Convolutional Neural Networks to Embedded Systems: The AlexNet and VGG-16 Case , 2018, 2018 17th ACM/IEEE International Conference on Information Processing in Sensor Networks (IPSN).

[24]  Lei Jiang,et al.  SHE: A Fast and Accurate Deep Neural Network for Encrypted Data , 2019, NeurIPS.

[25]  Yiran Chen,et al.  MoDNN: Local distributed mobile computing system for Deep Neural Network , 2017, Design, Automation & Test in Europe Conference & Exhibition (DATE), 2017.

[26]  Zahra Ghodsi,et al.  CryptoNAS: Private Inference on a ReLU Budget , 2020, NeurIPS.

[27]  Mohsen Guizani,et al.  DistPrivacy: Privacy-Aware Distributed Deep Neural Networks in IoT surveillance systems , 2020, GLOBECOM 2020 - 2020 IEEE Global Communications Conference.

[28]  Timo Aila,et al.  Pruning Convolutional Neural Networks for Resource Efficient Inference , 2016, ICLR.

[29]  Song Han,et al.  Deep Compression: Compressing Deep Neural Network with Pruning, Trained Quantization and Huffman Coding , 2015, ICLR.

[30]  Song Han,et al.  Learning both Weights and Connections for Efficient Neural Network , 2015, NIPS.