LFighter: Defending against the label-flipping attack in federated learning.

[1]  Shiping Wang,et al.  DBO-Net: Differentiable bi-level optimization network for multi-view clustering , 2023, Inf. Sci..

[2]  E. Wu,et al.  Two-phase Defense Against Poisoning Attacks on Federated Learning-based Intrusion Detection , 2023, Computers & security.

[3]  Alberto Blanco-Justicia,et al.  Enhanced Security and Privacy via Fragmented Federated Learning , 2022, IEEE Transactions on Neural Networks and Learning Systems.

[4]  J. Domingo-Ferrer,et al.  FL-Defender: Combating Targeted Attacks in Federated Learning , 2022, Knowl. Based Syst..

[5]  Muhammad Umar Janjua,et al.  Making federated learning robust to adversarial attacks by learning data and model association , 2022, Comput. Secur..

[6]  Alberto Blanco-Justicia,et al.  A Critical Review on the Use (and Misuse) of Differential Privacy in Machine Learning , 2022, ACM Comput. Surv..

[7]  J. Domingo-Ferrer,et al.  Explaining predictions and attacks in federated learning via random forests , 2022, Applied Intelligence.

[8]  Zhe Qu,et al.  LoMar: A Local Defense Against Poisoning Attack on Federated Learning , 2022, IEEE Transactions on Dependable and Secure Computing.

[9]  Josep Domingo-Ferrer,et al.  Secure and Privacy-Preserving Federated Learning via Co-Utility , 2021, IEEE Internet of Things Journal.

[10]  Dongcheng Li,et al.  Detection and Mitigation of Label-Flipping Attacks in Federated Learning Systems with KPCA and K-Means , 2021, 2021 8th International Conference on Dependable Systems and Their Applications (DSA).

[11]  Jian Lou,et al.  RobustFed: A Truth Inference Approach for Robust Federated Learning , 2021, CIKM.

[12]  Zhouchen Lin,et al.  Learning Deep Sparse Regularizers With Applications to Multi-View Clustering and Semi-Supervised Classification , 2021, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[13]  Josep Domingo-Ferrer,et al.  Achieving Security and Privacy in Federated Learning Systems: Survey, Research Challenges and Future Directions , 2020, Eng. Appl. Artif. Intell..

[14]  Mehmet Emre Gursoy,et al.  Data Poisoning Attacks Against Federated Learning Systems , 2020, ESORICS.

[15]  Kartik Sreenivasan,et al.  Attack of the Tails: Yes, You Really Can Backdoor Federated Learning , 2020, NeurIPS.

[16]  Josep Domingo-Ferrer,et al.  Co-Utile Peer-to-Peer Decentralized Computing , 2020, 2020 20th IEEE/ACM International Symposium on Cluster, Cloud and Internet Computing (CCGRID).

[17]  E. Cambria,et al.  Deep Learning--based Text Classification , 2020, ACM Comput. Surv..

[18]  G. Giannakis,et al.  Federated Variance-Reduced Stochastic Gradient Descent With Robustness to Byzantine Attacks , 2019, IEEE Transactions on Signal Processing.

[19]  Richard Nock,et al.  Advances and Open Problems in Federated Learning , 2019, Found. Trends Mach. Learn..

[20]  Sashank J. Reddi,et al.  SCAFFOLD: Stochastic Controlled Averaging for Federated Learning , 2019, ICML.

[21]  Amir Houmansadr,et al.  Comprehensive Privacy Analysis of Deep Learning: Passive and Active White-box Inference Attacks against Centralized and Federated Learning , 2018, 2019 IEEE Symposium on Security and Privacy (SP).

[22]  Nikita Borisov,et al.  Property Inference Attacks on Fully Connected Neural Networks using Permutation Invariant Representations , 2018, CCS.

[23]  Xu Chen,et al.  In-Edge AI: Intelligentizing Mobile Edge Computing, Caching and Communication by Federated Learning , 2018, IEEE Network.

[24]  Vitaly Shmatikov,et al.  How To Backdoor Federated Learning , 2018, AISTATS.

[25]  Vitaly Shmatikov,et al.  Exploiting Unintended Feature Leakage in Collaborative Learning , 2018, 2019 IEEE Symposium on Security and Privacy (SP).

[26]  Chang Liu,et al.  Manipulating Machine Learning: Poisoning Attacks and Countermeasures for Regression Learning , 2018, 2018 IEEE Symposium on Security and Privacy (SP).

[27]  Kannan Ramchandran,et al.  Byzantine-Robust Distributed Learning: Towards Optimal Statistical Rates , 2018, ICML.

[28]  Rachid Guerraoui,et al.  Machine Learning with Adversaries: Byzantine Tolerant Gradient Descent , 2017, NIPS.

[29]  Giuseppe Ateniese,et al.  Deep Models Under the GAN: Information Leakage from Collaborative Deep Learning , 2017, CCS.

[30]  Blaise Agüera y Arcas,et al.  Communication-Efficient Learning of Deep Networks from Decentralized Data , 2016, AISTATS.

[31]  Ricardo J. G. B. Campello,et al.  Density-Based Clustering Based on Hierarchical Density Estimates , 2013, PAKDD.

[32]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[33]  K. Esbensen,et al.  Principal component analysis , 1987 .

[34]  Geoffrey E. Hinton,et al.  Learning representations by back-propagating errors , 1986, Nature.

[35]  A. Siegel Robust regression using repeated medians , 1982 .

[36]  Zheli Liu,et al.  Differentially Private Byzantine-robust Federated Learning , 2022, IEEE Transactions on Parallel and Distributed Systems.

[37]  Fengjun Li,et al.  CONTRA: Defending Against Poisoning Attacks in Federated Learning , 2021, ESORICS.

[38]  Zhanghui Liu,et al.  Differentiable Bi-Sparse Multi-View Co-Clustering , 2021, IEEE Transactions on Signal Processing.

[39]  Somanath Tripathy,et al.  Moat: Model Agnostic Defense against Targeted Poisoning Attacks in Federated Learning , 2021, ICICS.

[40]  Josep Domingo-Ferrer,et al.  Efficient Detection of Byzantine Attacks in Federated Learning Using Last Layer Biases , 2020, MDAI.

[41]  Yoshua Bengio,et al.  Object Recognition with Gradient-Based Learning , 1999, Shape, Contour and Grouping in Computer Vision.

[42]  S. Roy,et al.  Factors influencing the choice of a learning rate for a backpropagation neural network , 1994, Proceedings of 1994 IEEE International Conference on Neural Networks (ICNN'94).