暂无分享,去创建一个
[1] Reza Shokri,et al. Comprehensive Privacy Analysis of Deep Learning: Stand-alone and Federated Learning under Passive and Active White-box Inference Attacks , 2018, ArXiv.
[2] Somesh Jha,et al. Privacy in Pharmacogenetics: An End-to-End Case Study of Personalized Warfarin Dosing , 2014, USENIX Security Symposium.
[3] Yoshua Bengio,et al. How transferable are features in deep neural networks? , 2014, NIPS.
[4] Haixu Tang,et al. Learning your identity and disease from research papers: information leaks in genome wide association study , 2009, CCS.
[5] Vitaly Shmatikov,et al. Membership Inference Attacks Against Machine Learning Models , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[6] Vitaly Shmatikov,et al. Privacy-preserving deep learning , 2015, 2015 53rd Annual Allerton Conference on Communication, Control, and Computing (Allerton).
[7] Ninghui Li,et al. Membership privacy: a unifying framework for privacy definitions , 2013, CCS.
[8] Ling Huang,et al. Learning in a Large Function Space: Privacy-Preserving Mechanisms for SVM Learning , 2009, J. Priv. Confidentiality.
[9] Prateek Jain,et al. To Drop or Not to Drop: Robustness, Consistency and Differential Privacy Properties of Dropout , 2015, ArXiv.
[10] Kai Chen,et al. Understanding Membership Inferences on Well-Generalized Learning Models , 2018, ArXiv.
[11] Vitaly Shmatikov,et al. The cost of privacy: destruction of data-mining utility in anonymized data publishing , 2008, KDD.
[12] Matt Fredrikson,et al. Influence-Directed Explanations for Deep Convolutional Networks , 2018, 2018 IEEE International Test Conference (ITC).
[13] H. Brendan McMahan,et al. A General Approach to Adding Differential Privacy to Iterative Training Procedures , 2018, ArXiv.
[14] Giovanni Felici,et al. Hacking smart machines with smarter ones: How to extract meaningful data from machine learning classifiers , 2013, Int. J. Secur. Networks.
[15] Ankur Taly,et al. Axiomatic Attribution for Deep Networks , 2017, ICML.
[16] Giuseppe Ateniese,et al. Deep Models Under the GAN: Information Leakage from Collaborative Deep Learning , 2017, CCS.
[17] Kevin P. Murphy,et al. Machine learning - a probabilistic perspective , 2012, Adaptive computation and machine learning series.
[18] S. Nelson,et al. Resolving Individuals Contributing Trace Amounts of DNA to Highly Complex Mixtures Using High-Density SNP Genotyping Microarrays , 2008, PLoS genetics.
[19] Yann LeCun,et al. The mnist database of handwritten digits , 2005 .
[20] Carl A. Gunter,et al. Towards Measuring Membership Privacy , 2017, ArXiv.
[21] Kilian Q. Weinberger,et al. On Calibration of Modern Neural Networks , 2017, ICML.
[22] Somesh Jha,et al. Model Inversion Attacks that Exploit Confidence Information and Basic Countermeasures , 2015, CCS.
[23] David Evans,et al. Evaluating Differentially Private Machine Learning in Practice , 2019, USENIX Security Symposium.
[24] Emiliano De Cristofaro,et al. LOGAN: Evaluating Privacy Leakage of Generative Models Using Generative Adversarial Networks , 2017, ArXiv.
[25] Pascal Vincent,et al. Unsupervised Feature Learning and Deep Learning: A Review and New Perspectives , 2012, ArXiv.
[26] Jeffrey F. Naughton,et al. A Methodology for Formalizing Model-Inversion Attacks , 2016, 2016 IEEE 29th Computer Security Foundations Symposium (CSF).
[27] Andrew Zisserman,et al. Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps , 2013, ICLR.
[28] Mario Fritz,et al. ML-Leaks: Model and Data Independent Membership Inference Attacks and Defenses on Machine Learning Models , 2018, NDSS.
[29] Michael I. Jordan,et al. Genomic privacy and limits of individual detection in a pool , 2009, Nature Genetics.
[30] Jun Tang,et al. Privacy Loss in Apple's Implementation of Differential Privacy on MacOS 10.12 , 2017, ArXiv.
[31] Fan Zhang,et al. Stealing Machine Learning Models via Prediction APIs , 2016, USENIX Security Symposium.
[32] Somesh Jha,et al. The Unintended Consequences of Overfitting: Training Data Inference Attacks , 2017, ArXiv.
[33] Jeffrey F. Naughton,et al. Revisiting Differentially Private Regression: Lessons From Learning Theory and their Consequences , 2015, ArXiv.
[34] Ian Goodfellow,et al. Deep Learning with Differential Privacy , 2016, CCS.
[35] Graham Cormode,et al. Personal privacy vs population privacy: learning to attack anonymization , 2011, KDD.
[36] Denis Nekipelov,et al. Estimation of Treatment Effects from Combined Data: Identification versus Data Security , 2015 .
[37] Cynthia Dwork,et al. Differential Privacy , 2006, ICALP.
[38] Guigang Zhang,et al. Deep Learning , 2016, Int. J. Semantic Comput..
[39] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[40] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[41] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[42] Eran Halperin,et al. Identifying Personal Genomes by Surname Inference , 2013, Science.
[43] Vitaly Shmatikov,et al. Inference Attacks Against Collaborative Learning , 2018, ArXiv.
[44] C. Bustamante,et al. Privacy Risks from Genomic Data-Sharing Beacons , 2015, American journal of human genetics.