Secure Bilevel Asynchronous Vertical Federated Learning with Backward Updating

Vertical federated learning (VFL) attracts increasing attention due to the emerging demands of multi-party collaborative modeling and concerns of privacy leakage. In the real VFL applications, usually only one or partial parties hold labels, which makes it challenging for all parties to collaboratively learn the model without privacy leakage. Meanwhile, most existing VFL algorithms are trapped in the synchronous computations, which leads to inefficiency in their real-world applications. To address these challenging problems, we propose a novel VFL framework integrated with new backward updating mechanism and bilevel asynchronous parallel architecture (VFB), under which three new algorithms, including VFB-SGD, SVRG, and -SAGA, are proposed. We derive the theoretical results of the convergence rates of these three algorithms under both strongly convex and nonconvex conditions. We also prove the security of VFB under semi-honest threat models. Extensive experiments on benchmark datasets demonstrate that our algorithms are efficient, scalable and lossless.

[1]  Richard Nock,et al.  Advances and Open Problems in Federated Learning , 2021, Found. Trends Mach. Learn..

[2]  Junchi Yan,et al.  Adversarial Learning for Robust Deep Clustering , 2020, NeurIPS.

[3]  Léon Bottou,et al.  Large-Scale Machine Learning with Stochastic Gradient Descent , 2010, COMPSTAT.

[4]  Moudud Alam,et al.  A Novel Generalized Ridge Regression Method for Quantitative Genetics , 2013, Genetics.

[5]  Xiaoqian Jiang,et al.  Secure Logistic Regression Based on Homomorphic Encryption: Design and Evaluation , 2018, IACR Cryptol. ePrint Arch..

[6]  Chih-Jen Lin,et al.  LIBSVM: A library for support vector machines , 2011, TIST.

[7]  Cheng Deng,et al.  Towards Transferable Targeted Attack , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[8]  Mariana Raykova,et al.  Secure Linear Regression on Vertically Partitioned Datasets , 2016, IACR Cryptol. ePrint Arch..

[9]  Chuishi Meng,et al.  Federated Forest , 2019, IEEE Transactions on Big Data.

[10]  Shiqian Ma,et al.  Stochastic Quasi-Newton Methods for Nonconvex Stochastic Optimization , 2014, SIAM J. Optim..

[11]  Xianglong Liu,et al.  Adversarial Fine-Grained Composition Learning for Unseen Attribute-Object Recognition , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).

[12]  J. Pei,et al.  Accelerated Zeroth-Order Momentum Methods from Mini to Minimax Optimization , 2020, arXiv.org.

[13]  Yijun Huang,et al.  Asynchronous Parallel Stochastic Gradient for Nonconvex Optimization , 2015, NIPS.

[14]  Richard Nock,et al.  Private federated learning on vertically partitioned data via entity resolution and additively homomorphic encryption , 2017, ArXiv.

[15]  Dacheng Tao,et al.  Learning Unseen Concepts via Hierarchical Decomposition and Composition , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[16]  Yuanming Shi,et al.  A Quasi-Newton Method Based Vertical Federated Learning Framework for Logistic Regression , 2019, ArXiv.

[17]  Yingbin Liang,et al.  SpiderBoost and Momentum: Faster Variance Reduction Algorithms , 2019, NeurIPS.

[18]  Xiang Li,et al.  Large-Scale Nonlinear AUC Maximization via Triply Stochastic Gradients. , 2020, IEEE transactions on pattern analysis and machine intelligence.

[19]  Cheng Deng,et al.  Faster Stochastic Quasi-Newton Methods. , 2021, IEEE transactions on neural networks and learning systems.

[20]  Yuguang Fang,et al.  Private Data Analytics on Biomedical Sensing Data via Distributed Computation , 2016, IEEE/ACM Transactions on Computational Biology and Bioinformatics.

[21]  Tianjian Chen,et al.  Federated Machine Learning: Concept and Applications , 2019 .

[22]  Runhua Xu,et al.  HybridAlpha: An Efficient Approach for Privacy-Preserving Federated Learning , 2019, AISec@CCS.

[23]  Paul Sajda,et al.  Fast, Exact Model Selection and Permutation Testing for l2-Regularized Logistic Regression , 2012, AISTATS.

[24]  Xiang Li,et al.  Federated Doubly Stochastic Kernel Learning for Vertically Partitioned Data , 2020, KDD.

[25]  Wei Zhang,et al.  Encrypted Gradient Descent Protocol for Outsourced Data Mining , 2015, 2015 IEEE 29th International Conference on Advanced Information Networking and Applications.

[26]  Tong Zhang,et al.  Accelerating Stochastic Gradient Descent using Predictive Variance Reduction , 2013, NIPS.

[27]  Fabian Pedregosa,et al.  ASAGA: Asynchronous Parallel SAGA , 2016, AISTATS.

[28]  Heng Huang,et al.  Faster Stochastic Alternating Direction Method of Multipliers for Nonconvex Optimization , 2020, ICML.

[29]  Ameet Talwalkar,et al.  Federated Multi-Task Learning , 2017, NIPS.

[30]  Cheng Deng,et al.  Privacy-Preserving Asynchronous Federated Learning Algorithms for Multi-Party Vertically Collaborative Learning , 2020, ArXiv.

[31]  Di Niu,et al.  FDML: A Collaborative Machine Learning Framework for Distributed Features , 2019, KDD.

[32]  Blaise Agüera y Arcas,et al.  Communication-Efficient Learning of Deep Networks from Decentralized Data , 2016, AISTATS.

[33]  Francis Bach,et al.  SAGA: A Fast Incremental Gradient Method With Support for Non-Strongly Convex Composite Objectives , 2014, NIPS.

[34]  Qiang Yang,et al.  SecureBoost: A Lossless Federated Learning Framework , 2019, IEEE Intelligent Systems.

[35]  Heng Huang,et al.  Asynchronous Mini-Batch Gradient Descent with Variance Reduction for Non-Convex Optimization , 2017, AAAI.

[36]  Hao Gao,et al.  Feature-Distributed SVRG for High-Dimensional Linear Classification , 2018, ArXiv.