DEAL: Decremental Energy-Aware Learning in a Federated System

Federated learning struggles with their heavy energy footprints on battery-powered devices. The learning process keeps all devices awake while draining expensive battery power to train a shared model collaboratively, yet it may still leak sensitive personal information. Traditional energy management techniques in system kernel mode can force the training device entering low power states, but it may violate the SLO of the collaborative learning. To address the conflict between learning SLO and energy efficiency, we propose DEAL, an energy efficient learning system that saves energy and preserves privacy with a decremental learning design. DEAL reduces the energy footprint from two layers: 1) an optimization layer that selects a subset of workers with sufficient capacity and maximum rewards. 2) a specified decremental learning algorithm that actively provides a decremental and incremental update functions, which allows kernel to correctly tune the local DVFS. We prototyped DEAL in containerized services with modern smartphone profiles and evaluated it with several learning benchmarks with realistic traces. We observed that DEAL achieves 75.6%–82.4% less energy footprint in different datasets, compared to the traditional methods. All learning processes are faster than state-of-thepractice FL frameworks up to 2-4X in model convergence.

[1]  Alexander J. Smola,et al.  Scaling Distributed Machine Learning with the Parameter Server , 2014, OSDI.

[2]  Li Li,et al.  Exploring federated learning on battery-powered devices , 2019, ACM TUR-C.

[3]  R. Cardell-Oliver,et al.  Dataset , 2019, Proceedings of the 2nd Workshop on Data Acquisition To Analysis - DATA'19.

[4]  Hamed Haddadi,et al.  Efficient and Private Federated Learning using TEE , 2019 .

[5]  Jie Zhou,et al.  CAPMAN: Cooling and Active Power Management in big.LITTLE Battery Supported Devices , 2020, 2020 IEEE 40th International Conference on Distributed Computing Systems (ICDCS).

[6]  Ameet Talwalkar,et al.  Federated Multi-Task Learning , 2017, NIPS.

[7]  Anusha Lalitha,et al.  Fully Decentralized Federated Learning , 2018 .

[8]  Volker Markl,et al.  Scalable similarity-based neighborhood methods with MapReduce , 2012, RecSys.

[9]  Robert H. Halstead,et al.  Matrix Computations , 2011, Encyclopedia of Parallel Computing.

[10]  Ali Dehghantanha,et al.  A survey on security and privacy of federated learning , 2021, Future Gener. Comput. Syst..

[11]  Sebastian Schelter "Amnesia" - Machine Learning Models That Can Forget User Data Very Fast , 2020, Conference on Innovative Data Systems Research.

[12]  Anit Kumar Sahu,et al.  Federated Learning: Challenges, Methods, and Future Directions , 2019, IEEE Signal Processing Magazine.

[13]  Prior Work , 2022 .

[14]  Yin Yang,et al.  FORA: Simple and Effective Approximate Single-Source Personalized PageRank , 2017, KDD.

[15]  C. W. Groetsch,et al.  The theory of Tikhonov regularization for Fredholm equations of the first kind , 1984 .

[16]  Marco Scavuzzo,et al.  Asynchronous Federated Learning for Geospatial Applications , 2018, DMLE/IOTSTREAMING@PKDD/ECML.

[17]  Carole-Jean Wu,et al.  Machine Learning at Facebook: Understanding Inference at the Edge , 2019, 2019 IEEE International Symposium on High Performance Computer Architecture (HPCA).

[18]  Hubert Eichner,et al.  Towards Federated Learning at Scale: System Design , 2019, SysML.

[19]  Peter Richtárik,et al.  Federated Learning: Strategies for Improving Communication Efficiency , 2016, ArXiv.

[20]  Christian M. Ernst,et al.  Multi-armed Bandit Allocation Indices , 1989 .

[21]  Michael J. Freedman,et al.  SLAQ: quality-driven scheduling for distributed machine learning , 2017, SoCC.

[22]  Richa Gupta,et al.  Smart Home With Virtual Assistant Using Raspberry Pi , 2019, 2019 9th International Conference on Cloud Computing, Data Science & Engineering (Confluence).

[23]  Blaise Agüera y Arcas,et al.  Communication-Efficient Learning of Deep Networks from Decentralized Data , 2016, AISTATS.

[24]  Ming Zhang,et al.  Where is the energy spent inside my app?: fine grained energy accounting on smartphones with Eprof , 2012, EuroSys '12.

[25]  Sarvar Patel,et al.  Practical Secure Aggregation for Privacy-Preserving Machine Learning , 2017, IACR Cryptol. ePrint Arch..

[26]  Bhushan Nemade Automatic Traffic Surveillance Using Video Tracking , 2016 .

[27]  Kin K. Leung,et al.  Adaptive Federated Learning in Resource Constrained Edge Computing Systems , 2018, IEEE Journal on Selected Areas in Communications.

[28]  Gert Cauwenberghs,et al.  Incremental and Decremental Support Vector Machine Learning , 2000, NIPS.

[29]  Manoj Kumar Tiwari,et al.  An efficient recommendation generation using relevant Jaccard similarity , 2019, Inf. Sci..

[30]  John Riedl,et al.  Item-based collaborative filtering recommendation algorithms , 2001, WWW '01.

[31]  Jie Xu,et al.  Course recommendation of MOOC with big data support: A contextual online learning approach , 2018, IEEE INFOCOM 2018 - IEEE Conference on Computer Communications Workshops (INFOCOM WKSHPS).

[32]  Jia Liu,et al.  Combinatorial Sleeping Bandits with Fairness Constraints , 2019, IEEE INFOCOM 2019 - IEEE Conference on Computer Communications.