NeuGuard: Lightweight Neuron-Guided Defense against Membership Inference Attacks
暂无分享,去创建一个
[1] Xiaoyong Yuan,et al. Membership Inference Attacks and Defenses in Neural Network Pruning , 2022, USENIX Security Symposium.
[2] Isamu Teranishi,et al. Knowledge Cross-Distillation for Membership Privacy , 2021, Proc. Priv. Enhancing Technol..
[3] Prateek Mittal,et al. Mitigating Membership Inference Attacks by Self-Distillation Through a Novel Ensemble Architecture , 2021, USENIX Security Symposium.
[4] Amir Houmansadr,et al. Membership Privacy for Machine Learning Models Through Knowledge Transfer , 2021, AAAI.
[5] N. Gong,et al. Practical Blind Membership Inference Attack via Differential Comparisons , 2021, NDSS.
[6] Yulun Zhang,et al. Neural Pruning via Growing Regularization , 2020, ICLR.
[7] Yang Zhang,et al. Membership Leakage in Label-Only Exposures , 2020, CCS.
[8] Nicolas Papernot,et al. Label-Only Membership Inference Attacks , 2020, ICML.
[9] Liwei Song,et al. Systematic Evaluation of Privacy Risks of Machine Learning Models , 2020, USENIX Security Symposium.
[10] Wenqi Wei,et al. Demystifying Membership Inference Attacks in Machine Learning as a Service , 2019, IEEE Transactions on Services Computing.
[11] Tudor Dumitras,et al. When Does Data Augmentation Help With Membership Inference Attacks? , 2021, ICML.
[12] Qijian He,et al. TransNet , 2020, Proc. VLDB Endow..
[13] Xiaochen Guo,et al. Stealing Your Data from Compressed Machine Learning Models , 2020, 2020 57th ACM/IEEE Design Automation Conference (DAC).
[14] Alex Orailoglu,et al. Concurrent Monitoring of Operational Health in Neural Networks Through Balanced Output Partitions , 2020, 2020 25th Asia and South Pacific Design Automation Conference (ASP-DAC).
[15] Mario Fritz,et al. GAN-Leaks: A Taxonomy of Membership Inference Attacks against Generative Models , 2019, CCS.
[16] Matt Fredrikson,et al. Stolen Memories: Leveraging Model Memorization for Calibrated White-Box Membership Inference , 2019, USENIX Security Symposium.
[17] Yang Zhang,et al. Updates-Leak: Data Set Inference and Reconstruction Attacks in Online Learning , 2019, USENIX Security Symposium.
[18] Yuheng Huang,et al. Neuron-level Structured Pruning using Polarization Regularizer , 2020, NeurIPS.
[19] Natalia Gimelshein,et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library , 2019, NeurIPS.
[20] Guojing Cong,et al. Accelerating Data Loading in Deep Neural Network Training , 2019, 2019 IEEE 26th International Conference on High Performance Computing, Data, and Analytics (HiPC).
[21] N. Gong,et al. MemGuard: Defending against Black-Box Membership Inference Attacks via Adversarial Examples , 2019, CCS.
[22] Wesley De Neve,et al. Impact of Adversarial Examples on Deep Learning Models for Biomedical Image Segmentation , 2019, MICCAI.
[23] Geoffrey E. Hinton,et al. When Does Label Smoothing Help? , 2019, NeurIPS.
[24] Prateek Mittal,et al. Privacy Risks of Securing Machine Learning Models against Adversarial Examples , 2019, CCS.
[25] Liwei Song,et al. Membership Inference Attacks Against Adversarially Robust Deep Learning Models , 2019, 2019 IEEE Security and Privacy Workshops (SPW).
[26] Dawn Song,et al. Towards Practical Differentially Private Convex Optimization , 2019, 2019 IEEE Symposium on Security and Privacy (SP).
[27] Calton Pu,et al. Differentially Private Model Publishing for Deep Learning , 2019, 2019 IEEE Symposium on Security and Privacy (SP).
[28] David Evans,et al. Evaluating Differentially Private Machine Learning in Practice , 2019, USENIX Security Symposium.
[29] Amir Houmansadr,et al. Comprehensive Privacy Analysis of Deep Learning: Passive and Active White-box Inference Attacks against Centralized and Federated Learning , 2018, 2019 IEEE Symposium on Security and Privacy (SP).
[30] Mario Fritz,et al. ML-Leaks: Model and Data Independent Membership Inference Attacks and Defenses on Machine Learning Models , 2018, NDSS.
[31] Bo Li,et al. Performing Co-membership Attacks Against Deep Generative Models , 2018, 2019 IEEE International Conference on Data Mining (ICDM).
[32] Vitaly Shmatikov,et al. Exploiting Unintended Feature Leakage in Collaborative Learning , 2018, 2019 IEEE Symposium on Security and Privacy (SP).
[33] Emiliano De Cristofaro,et al. LOGAN: Membership Inference Attacks Against Generative Models , 2017, Proc. Priv. Enhancing Technol..
[34] Ling Liu,et al. Towards Demystifying Membership Inference Attacks , 2018, ArXiv.
[35] Binghui Wang,et al. Stealing Hyperparameters in Machine Learning , 2018, 2018 IEEE Symposium on Security and Privacy (SP).
[36] David A. Wagner,et al. Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples , 2018, ICML.
[37] Reza Shokri,et al. Machine Learning with Membership Privacy using Adversarial Regularization , 2018, CCS.
[38] Somesh Jha,et al. Privacy Risk in Machine Learning: Analyzing the Connection to Overfitting , 2017, 2018 IEEE 31st Computer Security Foundations Symposium (CSF).
[39] Robert Laganière,et al. Membership Inference Attack against Differentially Private Deep Learning Model , 2018, Trans. Data Priv..
[40] Payman Mohassel,et al. SecureML: A System for Scalable Privacy-Preserving Machine Learning , 2017, 2017 IEEE Symposium on Security and Privacy (SP).
[41] Le Zhang,et al. AttriInfer: Inferring User Attributes in Online Social Networks Using Markov Random Fields , 2017, WWW.
[42] Nikos Komodakis,et al. Paying More Attention to Attention: Improving the Performance of Convolutional Neural Networks via Attention Transfer , 2016, ICLR.
[43] Samy Bengio,et al. Understanding deep learning requires rethinking generalization , 2016, ICLR.
[44] Vitaly Shmatikov,et al. Membership Inference Attacks Against Machine Learning Models , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[45] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[46] Fan Zhang,et al. Stealing Machine Learning Models via Prediction APIs , 2016, USENIX Security Symposium.
[47] Ian Goodfellow,et al. Deep Learning with Differential Privacy , 2016, CCS.
[48] Bin Liu,et al. You Are Who You Know and How You Behave: Attribute Inference Attacks via Users' Social Friends and Behaviors , 2016, USENIX Security Symposium.
[49] Sergey Ioffe,et al. Rethinking the Inception Architecture for Computer Vision , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[50] Somesh Jha,et al. Model Inversion Attacks that Exploit Confidence Information and Basic Countermeasures , 2015, CCS.
[51] Geoffrey E. Hinton,et al. Distilling the Knowledge in a Neural Network , 2015, ArXiv.
[52] Somesh Jha,et al. Privacy in Pharmacogenetics: An End-to-End Case Study of Personalized Warfarin Dosing , 2014, USENIX Security Symposium.
[53] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[54] Anand D. Sarwate,et al. Stochastic gradient descent with differentially private updates , 2013, 2013 IEEE Global Conference on Signal and Information Processing.
[55] V. Climenhaga. Markov chains and mixing times , 2013 .
[56] Geoffrey E. Hinton,et al. ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.
[57] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[58] Cynthia Dwork,et al. Calibrating Noise to Sensitivity in Private Data Analysis , 2006, TCC.
[59] John Tabak,et al. Geometry: The Language of Space and Form , 2004 .
[60] David J. C. MacKay,et al. Information Theory, Inference, and Learning Algorithms , 2004, IEEE Transactions on Information Theory.