The Privacy Onion Effect: Memorization is Relative
暂无分享,去创建一个
Florian Tramèr | Nicolas Papernot | Nicholas Carlini | A. Terzis | Matthew Jagielski | Chiyuan Zhang
[1] Florian Tramèr,et al. Truth Serum: Poisoning Machine Learning Models to Reveal Their Secrets , 2022, CCS.
[2] A. Madry,et al. Datamodels: Predicting Predictions from Training Data , 2022, ArXiv.
[3] Florian Tramèr,et al. Membership Inference Attacks From First Principles , 2021, 2022 IEEE Symposium on Security and Privacy (SP).
[4] Nicolas Papernot,et al. Unrolling SGD: Understanding Factors Influencing Machine Unlearning , 2021, 2022 IEEE 7th European Symposium on Security and Privacy (EuroS&P).
[5] M. Zeppelzauer,et al. Machine unlearning: linear filtration for logit-based classifiers , 2020, Machine Learning.
[6] Daphne Ippolito,et al. Counterfactual Memorization in Neural Language Models , 2021, ArXiv.
[7] Zhumin Chen,et al. Membership Inference Attacks Against Recommender Systems , 2021, CCS.
[8] Neil Zhenqiang Gong,et al. EncoderMI: Membership Inference against Pre-trained Encoders in Contrastive Learning , 2021, CCS.
[9] Jonas Mueller,et al. Pervasive Label Errors in Test Sets Destabilize Machine Learning Benchmarks , 2021, NeurIPS Datasets and Benchmarks.
[10] Samy Bengio,et al. Understanding deep learning (still) requires rethinking generalization , 2021, Commun. ACM.
[11] Colin Raffel,et al. Extracting Training Data from Large Language Models , 2020, USENIX Security Symposium.
[12] Vitaly Feldman,et al. When is memorization of irrelevant training data necessary for high-accuracy learning? , 2020, STOC.
[13] Dan Boneh,et al. Differentially Private Learning Needs Better Features (or Much More Data) , 2020, ICLR.
[14] Vijay Ganesh,et al. Amnesiac Machine Learning , 2020, AAAI.
[15] Marzyeh Ghassemi,et al. Chasing Your Long Tails: Differentially Private Prediction in Health Care Settings , 2020, FAccT.
[16] Nicolas Papernot,et al. Label-Only Membership Inference Attacks , 2020, ICML.
[17] Ziheng Jiang,et al. Characterizing Structural Regularities of Labeled Data in Overparameterized Models , 2020, ICML.
[18] Vitaly Feldman,et al. What Neural Networks Memorize and Why: Discovering the Long Tail via Influence Estimation , 2020, NeurIPS.
[19] Reza Shokri,et al. ML Privacy Meter: Aiding Regulatory Compliance by Quantifying the Privacy Risks of Machine Learning , 2020, ArXiv.
[20] Jonathan Ullman,et al. Auditing Differentially Private Machine Learning: How Private is Private SGD? , 2020, NeurIPS.
[21] Matt Fredrikson,et al. Stolen Memories: Leveraging Model Memorization for Calibrated White-Box Membership Inference , 2019, USENIX Security Symposium.
[22] Vitaly Feldman,et al. Does learning require memorization? a short tale about a long tail , 2019, STOC.
[23] Vitaly Shmatikov,et al. Differential Privacy Has Disparate Impact on Model Accuracy , 2019, NeurIPS.
[24] Cordelia Schmid,et al. White-box vs Black-box: Bayes Optimal Strategies for Membership Inference , 2019, ICML.
[25] Úlfar Erlingsson,et al. The Secret Sharer: Evaluating and Testing Unintended Memorization in Neural Networks , 2018, USENIX Security Symposium.
[26] Reza Shokri,et al. Comprehensive Privacy Analysis of Deep Learning: Stand-alone and Federated Learning under Passive and Active White-box Inference Attacks , 2018, ArXiv.
[27] Somesh Jha,et al. Privacy Risk in Machine Learning: Analyzing the Connection to Overfitting , 2017, 2018 IEEE 31st Computer Security Foundations Symposium (CSF).
[28] Emiliano De Cristofaro,et al. Knock Knock, Who's There? Membership Inference on Aggregate Location Data , 2017, NDSS.
[29] Vitaly Shmatikov,et al. Membership Inference Attacks Against Machine Learning Models , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[30] Ian Goodfellow,et al. Deep Learning with Differential Privacy , 2016, CCS.
[31] Junfeng Yang,et al. Towards Making Systems Forget with Machine Unlearning , 2015, 2015 IEEE Symposium on Security and Privacy.
[32] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[33] S. Nelson,et al. Resolving Individuals Contributing Trace Amounts of DNA to Highly Complex Mixtures Using High-Density SNP Genotyping Microarrays , 2008, PLoS genetics.
[34] Sofya Raskhodnikova,et al. Smooth sensitivity and sampling in private data analysis , 2007, STOC '07.
[35] Cynthia Dwork,et al. Calibrating Noise to Sensitivity in Private Data Analysis , 2006, TCC.