Incentive-Aware Decentralized Data Collaboration

Data collaboration enables multiple parties to pool data for deriving meaningful data insights. However, data misuse and unlawful data collection have led to precautionary measures being imposed by individual organizations to guide against data leakage and abuse. As a response, decentralized federated learning (DFL) has emerged as an attractive paradigm to facilitate data collaboration while being amenable to privacy-preserving data and knowledge sharing, cost reduction, and prediction accuracy improvement. Unfortunately, the participating parties in DFL tend to be heterogeneous with skew datasets and uneven capabilities. Inevitably, training and transmission costs, and the presence of free-riders pose challenges to the adoption and participation of DFL. The absence of centralized parameter servers further exacerbates the problem of evaluating the contribution of each individual party. Therefore, an effective incentive mechanism is essential to promote data collaboration. In this paper, we propose a novel Incentive-aware Decentralized fEderated leArning (IDEA) framework for facilitating data collaboration. Specifically, we first design a customizable reward scheme for heterogeneous parties to optimize their respective objectives such as higher model accuracy, communication efficiency, and computational efficiency. To reward fairly to deserving parties while offering flexibility, we propose a novel multi-agent reinforcement learning (MARL) incentive mechanism, which enables heterogeneous parties to learn their own optimal collaboration policy. We then design an efficient decentralized data collaboration algorithm that supports the customizable reward scheme based on individual objective-specific collaboration policy. We theoretically prove that the algorithm achieves a Nash equilibrium, which ensures the fairness of the corresponding rewards for parties. We conduct extensive experiments to evaluate the performance of our proposed framework against four baselines on five real-world datasets. The results show that IDEA outperforms state-of-the-art methods in terms of effectiveness, efficiency, and accumulated reward.

[1]  Khin Mi Mi Aung,et al.  Skellam Mixture Mechanism: a Novel Approach to Federated Learning with Differential Privacy , 2022, Proc. VLDB Endow..

[2]  B. Ooi,et al.  DyHealth , 2022, Proceedings of the VLDB Endowment.

[3]  Khin Mi Mi Aung,et al.  Skellam mixture mechanism , 2022, Proceedings of the VLDB Endowment.

[4]  Mike Zheng Shou,et al.  The Metaverse Data Deluge: What Can We Do About It? , 2022, 2023 IEEE 39th International Conference on Data Engineering (ICDE).

[5]  Yong Cheng,et al.  BlindFL: Vertical Federated Machine Learning without Peeking into Your Data , 2022, SIGMOD Conference.

[6]  Dusit Niyato,et al.  Incentive Mechanisms for Federated Learning: From Economic and Game Theoretic Perspective , 2021, IEEE Transactions on Cognitive Communications and Networking.

[7]  Octavia A. Dobre,et al.  Federated Learning for Smart Healthcare: A Survey , 2021, ACM Comput. Surv..

[8]  Venkatesh Saligrama,et al.  Federated Learning Based on Dynamic Regularization , 2021, ICLR.

[9]  M. Friedlander,et al.  Improving Fairness for Data Valuation in Horizontal Federated Learning , 2021, 2022 IEEE 38th International Conference on Data Engineering (ICDE).

[10]  Eugene Wu,et al.  Enabling SQL-based Training Data Debugging for Federated Learning , 2021, Proc. VLDB Endow..

[11]  Olof Mogren,et al.  Decentralized federated learning of deep neural networks on non-iid data , 2021, ArXiv.

[12]  Luis Muñoz-González,et al.  Non-IID data re-balancing at IoT edge with peer-to-peer federated learning for anomaly detection , 2021, WISEC.

[13]  Jiawei Jiang,et al.  VF2Boost: Very Fast Vertical Federated Gradient Boosting for Cross-Enterprise Learning , 2021, SIGMOD Conference.

[14]  Ali Dehghantanha,et al.  Federated-Learning-Based Anomaly Detection for IoT Security Attacks , 2021, IEEE Internet of Things Journal.

[15]  N. Gatti,et al.  Simple Uncoupled No-regret Learning Dynamics for Extensive-form Correlated Equilibrium , 2021, J. ACM.

[16]  Bingsheng He,et al.  Federated Learning on Non-IID Data Silos: An Experimental Study , 2021, 2022 IEEE 38th International Conference on Data Engineering (ICDE).

[17]  Shui Yu,et al.  FASTGNN: A Topological Information Protected Federated Learning Approach for Traffic Speed Forecasting , 2021, IEEE Transactions on Industrial Informatics.

[18]  Yang Cao,et al.  Transparent Contribution Evaluation for Secure Federated Learning on Blockchain , 2021, 2021 IEEE 37th International Conference on Data Engineering Workshops (ICDEW).

[19]  Pramod K. Varshney,et al.  Decentralized Federated Learning via Mutual Knowledge Transfer , 2020, IEEE Internet of Things Journal.

[20]  Cong Wang,et al.  FedServing: A Federated Prediction Serving Framework Based on Incentive Mechanism , 2020, IEEE INFOCOM 2021 - IEEE Conference on Computer Communications.

[21]  Bingbing Ni,et al.  MedMNIST Classification Decathlon: A Lightweight AutoML Benchmark for Medical Image Analysis , 2020, 2021 IEEE 18th International Symposium on Biomedical Imaging (ISBI).

[22]  Rong-Rong Chen,et al.  Local Averaging Helps: Hierarchical Federated Learning and Convergence Analysis , 2020, ArXiv.

[23]  Beng Chin Ooi,et al.  Feature Inference Attack on Model Predictions in Vertical Federated Learning , 2020, 2021 IEEE 37th International Conference on Data Engineering (ICDE).

[24]  B. Ooi,et al.  Communication-efficient Decentralized Machine Learning over Heterogeneous Networks , 2020, 2021 IEEE 37th International Conference on Data Engineering (ICDE).

[25]  Mun Choon Chan,et al.  Collaborative Machine Learning with Incentive-Aware Model Rewards , 2020, ICML.

[26]  Shuguang Cui,et al.  Delay Minimization for Federated Learning Over Wireless Communication Networks , 2020, ArXiv.

[27]  Beng Chin Ooi,et al.  Privacy preserving vertical federated learning for tree-based models , 2020, Proc. VLDB Endow..

[28]  Khaled Salah,et al.  Towards Blockchain-Based Reputation-Aware Federated Learning , 2020, IEEE INFOCOM 2020 - IEEE Conference on Computer Communications Workshops (INFOCOM WKSHPS).

[29]  D. Jannach,et al.  Federated Meta-Learning for Fraudulent Credit Card Detection , 2020, IJCAI.

[30]  Yang Liu,et al.  A Sustainable Incentive Scheme for Federated Learning , 2020, IEEE Intelligent Systems.

[31]  Richard Vidal,et al.  Free-rider Attacks on Model Aggregation in Federated Learning , 2020, AISTATS.

[32]  Monica Nicoli,et al.  Federated Learning with Mutually Cooperating Devices: A Consensus Approach Towards Server-Less Model Optimization , 2020, ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[33]  Christopher Briggs,et al.  Federated learning with hierarchical clustering of local updates to improve training on non-IID data , 2020, 2020 International Joint Conference on Neural Networks (IJCNN).

[34]  Chunyan Miao,et al.  Hierarchical Incentive Mechanism Design for Federated Machine Learning in Mobile Networks , 2020, IEEE Internet of Things Journal.

[35]  Michael Moeller,et al.  Inverting Gradients - How easy is it to break privacy in federated learning? , 2020, NeurIPS.

[36]  Micah J. Sheller,et al.  The future of digital health with federated learning , 2020, npj Digital Medicine.

[37]  Yunlong Wang,et al.  Decentralized Federated Learning for Electronic Health Records , 2020, 2020 54th Annual Conference on Information Sciences and Systems (CISS).

[38]  X. Chu,et al.  FMore: An Incentive Scheme of Multi-dimensional Auction for Federated Learning in MEC , 2020, 2020 IEEE 40th International Conference on Distributed Computing Systems (ICDCS).

[39]  Deze Zeng,et al.  A Learning-Based Incentive Mechanism for Federated Learning , 2020, IEEE Internet of Things Journal.

[40]  Monica Nicoli,et al.  Federated Learning With Cooperating Devices: A Consensus Approach for Massive IoT Networks , 2019, IEEE Internet of Things Journal.

[41]  Zhu Han,et al.  Incentivize to Build: A Crowdsourcing Framework for Federated Learning , 2019, 2019 IEEE Global Communications Conference (GLOBECOM).

[42]  T. Başar,et al.  Multi-Agent Reinforcement Learning: A Selective Overview of Theories and Algorithms , 2019, Handbook of Reinforcement Learning and Control.

[43]  Mohsen Guizani,et al.  Reliable Federated Learning for Mobile Networks , 2019, IEEE Wireless Communications.

[44]  Yuanqing Xia,et al.  Crowdtraining: Architecture and Incentive Mechanism for Deep Learning Training in the Internet of Things , 2019, IEEE Network.

[45]  Anit Kumar Sahu,et al.  Federated Learning: Challenges, Methods, and Future Directions , 2019, IEEE Signal Processing Magazine.

[46]  Bing Chen,et al.  Poisoning Attack in Federated Learning using Generative Adversarial Nets , 2019, 2019 18th IEEE International Conference On Trust, Security And Privacy In Computing And Communications/13th IEEE International Conference On Big Data Science And Engineering (TrustCom/BigDataSE).

[47]  István Hegedüs,et al.  Gossip Learning as a Decentralized Alternative to Federated Learning , 2019, DAIS.

[48]  Yasaman Khazaeni,et al.  Bayesian Nonparametric Federated Learning of Neural Networks , 2019, ICML.

[49]  Nassir Navab,et al.  BrainTorrent: A Peer-to-Peer Environment for Decentralized Federated Learning , 2019, ArXiv.

[50]  Martin Jaggi,et al.  Decentralized Stochastic Optimization and Gossip Algorithms with Compressed Communication , 2019, ICML.

[51]  Yuanzhou Yang,et al.  Highly Scalable Deep Learning Training System with Mixed-Precision: Training ImageNet in Four Minutes , 2018, ArXiv.

[52]  Paul Voigt,et al.  The Eu General Data Protection Regulation (Gdpr): A Practical Guide , 2017 .

[53]  Chinmay Hegde,et al.  Collaborative Deep Learning in Fixed Topology Networks , 2017, NIPS.

[54]  Wei Zhang,et al.  Can Decentralized Algorithms Outperform Centralized Algorithms? A Case Study for Decentralized Parallel Stochastic Gradient Descent , 2017, NIPS.

[55]  Anthony K. H. Tung,et al.  SINGA: A Distributed Deep Learning Platform , 2015, ACM Multimedia.

[56]  Andrew Zisserman,et al.  Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.

[57]  Michael P. Wellman,et al.  Nash Q-Learning for General-Sum Stochastic Games , 2003, J. Mach. Learn. Res..

[58]  A. Barabasi,et al.  Mean-field theory for scale-free random networks , 1999, cond-mat/9907068.

[59]  Avi Pfeffer,et al.  Representations and Solutions for Game-Theoretic Problems , 1997, Artif. Intell..

[60]  Gerald Tesauro,et al.  Temporal difference learning and TD-Gammon , 1995, CACM.

[61]  L. Blume The Statistical Mechanics of Strategic Interaction , 1993 .

[62]  Kaiping Zheng DyHealth: Making Neural Networks Dynamic for Effective Healthcare Analytics , 2022, Proc. VLDB Endow..

[63]  Gauri Joshi,et al.  Cooperative SGD: A Unified Framework for the Design and Analysis of Local-Update SGD Algorithms , 2021, J. Mach. Learn. Res..

[64]  Lidan Shou,et al.  Refiner: A Reliable Incentive-Driven Federated Learning System Powered by Blockchain , 2021, Proc. VLDB Endow..

[65]  H. Young,et al.  Handbook of Game Theory with Economic Applications , 2015 .

[66]  Alex Krizhevsky,et al.  Learning Multiple Layers of Features from Tiny Images , 2009 .

[67]  A. Barabasi,et al.  Emergence of Scaling in Random Networks , 1999 .

[68]  Michael I. Jordan,et al.  On the Convergence of Stochastic Iterative Dynamic Programming Algorithms , 1994, Neural Computation.

[69]  Sergiu Hart,et al.  Games in extensive and strategic forms , 1992 .

[70]  A. M. Fink,et al.  Equilibrium in a stochastic $n$-person game , 1964 .

[71]  広島大学 Journal of science of the Hiroshima University. Ser. A-1, Mathematics , 1961 .