暂无分享,去创建一个
Huseyin A. Inan | Osman Ramadan | Robert Sim | Daniel Jones | James Withers | Lukas Wutschitz | Victor Rühle | Daniel Jones | L. Wutschitz | Victor Rühle | Robert Sim | James Withers | Osman Ramadan | Lukas Wutschitz
[1] Yu Sun,et al. ERNIE: Enhanced Representation through Knowledge Integration , 2019, ArXiv.
[2] Thorsten Brants,et al. One billion word benchmark for measuring progress in statistical language modeling , 2013, INTERSPEECH.
[3] Matt Fredrikson,et al. Stolen Memories: Leveraging Model Memorization for Calibrated White-Box Membership Inference , 2019, USENIX Security Symposium.
[4] Ling Liu,et al. Towards Demystifying Membership Inference Attacks , 2018, ArXiv.
[5] Vitaly Shmatikov,et al. Membership Inference Attacks Against Machine Learning Models , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[6] Colin Raffel,et al. Extracting Training Data from Large Language Models , 2020, USENIX Security Symposium.
[7] Vitaly Shmatikov,et al. Auditing Data Provenance in Text-Generation Models , 2018, KDD.
[8] Anand D. Sarwate,et al. Stochastic gradient descent with differentially private updates , 2013, 2013 IEEE Global Conference on Signal and Information Processing.
[9] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[10] Vitaly Feldman,et al. Does learning require memorization? a short tale about a long tail , 2019, STOC.
[11] Kevin Duh,et al. Membership Inference Attacks on Sequence-to-Sequence Models: Is My Data In Your Machine Translation System? , 2019, TACL.
[12] Yael Kovo. Office of Science and Technology Policy (OSTP) , 2016 .
[13] Santiago Zanella Béguelin,et al. Analyzing Information Leakage of Updates to Natural Language Models , 2019, CCS.
[14] Úlfar Erlingsson,et al. The Secret Sharer: Evaluating and Testing Unintended Memorization in Neural Networks , 2018, USENIX Security Symposium.
[15] Sebastian Ruder,et al. Universal Language Model Fine-tuning for Text Classification , 2018, ACL.
[16] Andrew M. Dai,et al. Gmail Smart Compose: Real-Time Assisted Writing , 2019, KDD.
[17] M. Newman. Power laws, Pareto distributions and Zipf's law , 2005 .
[18] Somesh Jha,et al. Privacy Risk in Machine Learning: Analyzing the Connection to Overfitting , 2017, 2018 IEEE 31st Computer Security Foundations Symposium (CSF).
[19] Paul Irolla,et al. Demystifying the Membership Inference Attack , 2019, 2019 12th CMI Conference on Cybersecurity and Privacy (CMI).
[20] Amir Houmansadr,et al. Comprehensive Privacy Analysis of Deep Learning: Passive and Active White-box Inference Attacks against Centralized and Federated Learning , 2018, 2019 IEEE Symposium on Security and Privacy (SP).
[21] Ian Goodfellow,et al. Deep Learning with Differential Privacy , 2016, CCS.
[22] Sebastian Riedel,et al. Language Models as Knowledge Bases? , 2019, EMNLP.
[23] Dan Boneh,et al. Differentially Private Learning Needs Better Features (or Much More Data) , 2020, ICLR.
[24] Nicholas Carlini,et al. Label-Only Membership Inference Attacks , 2020, ICML.
[25] Hermann Ney,et al. LSTM Neural Networks for Language Modeling , 2012, INTERSPEECH.
[26] Emiliano De Cristofaro,et al. LOGAN: Membership Inference Attacks Against Generative Models , 2017, Proc. Priv. Enhancing Technol..
[27] Alec Radford,et al. Improving Language Understanding by Generative Pre-Training , 2018 .
[28] Yiming Yang,et al. XLNet: Generalized Autoregressive Pretraining for Language Understanding , 2019, NeurIPS.
[29] Latanya Sweeney,et al. k-Anonymity: A Model for Protecting Privacy , 2002, Int. J. Uncertain. Fuzziness Knowl. Based Syst..
[30] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[31] Cordelia Schmid,et al. White-box vs Black-box: Bayes Optimal Strategies for Membership Inference , 2019, ICML.
[32] Mario Fritz,et al. ML-Leaks: Model and Data Independent Membership Inference Attacks and Defenses on Machine Learning Models , 2018, NDSS.
[33] Mark Chen,et al. Language Models are Few-Shot Learners , 2020, NeurIPS.
[34] Javier Snaider,et al. Conversational Contextual Cues: The Case of Personalization and History for Response Ranking , 2016, ArXiv.
[35] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[36] Kai Chen,et al. Understanding Membership Inferences on Well-Generalized Learning Models , 2018, ArXiv.
[37] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.
[38] Samy Bengio,et al. Understanding deep learning requires rethinking generalization , 2016, ICLR.
[39] Ilya Sutskever,et al. Language Models are Unsupervised Multitask Learners , 2019 .
[40] Yoshua Bengio,et al. A Neural Probabilistic Language Model , 2003, J. Mach. Learn. Res..
[41] Lukás Burget,et al. Recurrent neural network based language model , 2010, INTERSPEECH.
[42] Reza Shokri,et al. ML Privacy Meter: Aiding Regulatory Compliance by Quantifying the Privacy Risks of Machine Learning , 2020, ArXiv.
[43] Vitaly Feldman,et al. When is memorization of irrelevant training data necessary for high-accuracy learning? , 2020, STOC.
[44] Shuang Song,et al. Making the Shoe Fit: Architectures, Initializations, and Tuning for Learning with Privacy , 2019 .