Data Shuffling in Wireless Distributed Computing via Low-Rank Optimization

Intelligent mobile platforms such as smart vehicles and drones have recently become the focus of attention for onboard deployment of machine learning mechanisms to enable low latency decisions with low risk of privacy breach. However, most such machine learning algorithms are both computation-and-memory intensive, which makes it highly difficult to implement the requisite computations on a single device of limited computation, memory, and energy resources. Wireless distributed computing presents new opportunities by pooling the computation and storage resources among devices. For low-latency applications, the key bottleneck lies in the exchange of intermediate results among mobile devices for data shuffling. To improve communication efficiency, we propose a co-channel communication model and design transceivers by exploiting the locally computed intermediate values as side information. A low-rank optimization model is proposed to maximize the achieved degrees-of-freedom (DoF) by establishing the interference alignment condition for data shuffling. Unfortunately, existing approaches to approximate the rank function fail to yield satisfactory performance due to the poor structure in the formulated low-rank optimization problem. In this paper, we develop an efficient difference-of-convex-functions (DC) algorithm to solve the presented low-rank optimization problem by proposing a novel DC representation for the rank function. Numerical experiments demonstrate that the proposed DC approach can significantly improve the communication efficiency whereas the achievable DoF almost remains unchanged when the number of mobile devices grows.

[1]  T. P. Dinh,et al.  Convex analysis approach to d.c. programming: Theory, Algorithm and Applications , 1997 .

[2]  William J. Dally,et al.  Deep Gradient Compression: Reducing the Communication Bandwidth for Distributed Training , 2017, ICLR.

[3]  Sanjay Ghemawat,et al.  MapReduce: Simplified Data Processing on Large Clusters , 2004, OSDI.

[4]  Yuanming Shi,et al.  Large-Scale Convex Optimization for Dense Wireless Cooperative Networks , 2015, IEEE Transactions on Signal Processing.

[5]  Akiko Takeda,et al.  DC formulations and algorithms for sparse optimization problems , 2017, Mathematical Programming.

[6]  Tao Zhang,et al.  Model Compression and Acceleration for Deep Neural Networks: The Principles, Progress, and Challenges , 2018, IEEE Signal Processing Magazine.

[7]  Xuan Vinh Doan,et al.  Finding the Largest Low-Rank Clusters With Ky Fan 2-k-Norm and ℓ1-Norm , 2014, SIAM J. Optim..

[8]  A. Salman Avestimehr,et al.  A Fundamental Tradeoff Between Computation and Communication in Distributed Computing , 2016, IEEE Transactions on Information Theory.

[9]  Justin K. Romberg,et al.  An Overview of Low-Rank Matrix Recovery From Incomplete Observations , 2016, IEEE Journal of Selected Topics in Signal Processing.

[10]  A. Salman Avestimehr,et al.  A Scalable Framework for Wireless Distributed Computing , 2016, IEEE/ACM Transactions on Networking.

[11]  Le Thi Hoai An,et al.  DC programming and DCA: thirty years of developments , 2018, Math. Program..

[12]  G. Alistair Watson,et al.  On matrix approximation problems with Ky Fank norms , 1993, Numerical Algorithms.

[13]  Maryam Fazel,et al.  Iterative reweighted algorithms for matrix rank minimization , 2012, J. Mach. Learn. Res..

[14]  丸山 徹 Convex Analysisの二,三の進展について , 1977 .

[15]  Stephen P. Boyd,et al.  Convex Optimization , 2004, Algorithms and Theory of Computation Handbook.

[16]  Zhi Ding,et al.  Low-Rank Optimization for Data Shuffling in Wireless Distributed Computing , 2018, 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[17]  Sergios Theodoridis,et al.  Adaptive Learning in Complex Reproducing Kernel Hilbert Spaces Employing Wirtinger's Subgradients , 2012, IEEE Transactions on Neural Networks and Learning Systems.

[18]  G. Watson Characterization of the subdifferential of some matrix norms , 1992 .

[19]  Yuanming Shi,et al.  Low-Rank Matrix Completion for Topological Interference Management by Riemannian Pursuit , 2016, IEEE Transactions on Wireless Communications.

[20]  Syed Ali Jafar,et al.  Index Coding - An Interference Alignment Perspective , 2014, IEEE Trans. Inf. Theory.

[21]  Syed Ali Jafar,et al.  Interference Alignment and Degrees of Freedom of the $K$-User Interference Channel , 2008, IEEE Transactions on Information Theory.

[22]  David Tse,et al.  Feasibility of Interference Alignment for the MIMO Interference Channel , 2013, IEEE Transactions on Information Theory.

[23]  Vivienne Sze,et al.  Efficient Processing of Deep Neural Networks: A Tutorial and Survey , 2017, Proceedings of the IEEE.

[24]  Song Han,et al.  Deep Compression: Compressing Deep Neural Network with Pruning, Trained Quantization and Huffman Coding , 2015, ICLR.

[25]  Thomas M. Cover,et al.  Elements of Information Theory , 2005 .

[26]  Stephen P. Boyd,et al.  Conic Optimization via Operator Splitting and Homogeneous Self-Dual Embedding , 2013, Journal of Optimization Theory and Applications.