Distilling On-Device Intelligence at the Network Edge

Devices at the edge of wireless networks are the last mile data sources for machine learning (ML). As opposed to traditional ready-made public datasets, these user-generated private datasets reflect the freshest local environments in real time. They are thus indispensable for enabling mission-critical intelligent systems, ranging from fog radio access networks (RANs) to driverless cars and e-Health wearables. This article focuses on how to distill high-quality on-device ML models using fog computing, from such user-generated private data dispersed across wirelessly connected devices. To this end, we introduce communication-efficient and privacy-preserving distributed ML frameworks, termed fog ML (FML), wherein on-device ML models are trained by exchanging model parameters, model outputs, and surrogate data. We then present advanced FML frameworks addressing wireless RAN characteristics, limited on-device resources, and imbalanced data distributions. Our study suggests that the full potential of FML can be reached by co-designing communication and distributed ML operations while accounting for heterogeneous hardware specifications, data characteristics, and user requirements.

[1]  Blaise Agüera y Arcas,et al.  Communication-Efficient Learning of Deep Networks from Decentralized Data , 2016, AISTATS.

[2]  Mehdi Bennis,et al.  Communication-Efficient On-Device Machine Learning: Federated Distillation and Augmentation under Non-IID Private Data , 2018, ArXiv.

[3]  Han Cha,et al.  Federated Reinforcement Distillation with Proxy Experience Memory , 2019, ArXiv.

[4]  Seong-Lyun Kim,et al.  Tractable Resource Management With Uplink Decoupled Millimeter-Wave Overlay in Ultra-Dense Cellular Networks , 2015, IEEE Transactions on Wireless Communications.

[5]  Seong-Lyun Kim,et al.  Blockchained On-Device Federated Learning , 2018, IEEE Communications Letters.

[6]  Ramesh Raskar,et al.  No Peek: A Survey of private distributed deep learning , 2018, ArXiv.

[7]  Seong-Lyun Kim,et al.  Multi-hop Federated Private Data Augmentation with Sample Compression , 2019, ArXiv.

[8]  Mehdi Bennis,et al.  Wireless Network Intelligence at the Edge , 2018, Proceedings of the IEEE.

[9]  Geoffrey E. Hinton,et al.  Large scale distributed neural network training through online distillation , 2018, ICLR.

[10]  Peng Jiang,et al.  A Linear Speedup Analysis of Distributed Deep Learning with Sparse and Quantized Communication , 2018, NeurIPS.

[11]  Kin K. Leung,et al.  Adaptive Federated Learning in Resource Constrained Edge Computing Systems , 2018, IEEE Journal on Selected Areas in Communications.

[12]  Qiang Yang,et al.  Federated Machine Learning , 2019, ACM Trans. Intell. Syst. Technol..

[13]  Mehdi Bennis,et al.  GADMM: Fast and Communication Efficient Framework for Distributed Machine Learning , 2019, J. Mach. Learn. Res..

[14]  Ting He,et al.  Robust Coreset Construction for Distributed Machine Learning , 2019, 2019 IEEE Global Communications Conference (GLOBECOM).

[15]  Shiqiang Wang,et al.  On Data Summarization for Machine Learning in Multi-organization Federations , 2019, 2019 IEEE International Conference on Smart Computing (SMARTCOMP).