暂无分享,去创建一个
Peter Kairouz | Lalitha Sankar | Jiachun Liao | Mario Diaz | P. Kairouz | Mario Díaz | Jiachun Liao | L. Sankar
[1] G. Lewicki,et al. Approximation by Superpositions of a Sigmoidal Function , 2003 .
[2] Kurt Hornik,et al. Multilayer feedforward networks are universal approximators , 1989, Neural Networks.
[3] Vitaly Shmatikov,et al. Exploiting Unintended Feature Leakage in Collaborative Learning , 2018, 2019 IEEE Symposium on Security and Privacy (SP).
[4] Ohad Shamir,et al. Learnability, Stability and Uniform Convergence , 2010, J. Mach. Learn. Res..
[5] Somesh Jha,et al. Model Inversion Attacks that Exploit Confidence Information and Basic Countermeasures , 2015, CCS.
[6] Ram Rajagopal,et al. Generative Adversarial Privacy: A Data-Driven Approach to Information-Theoretic Privacy , 2018, 2018 52nd Asilomar Conference on Signals, Systems, and Computers.
[7] David Evans,et al. Evaluating Differentially Private Machine Learning in Practice , 2019, USENIX Security Symposium.
[8] James Zou,et al. Controlling Bias in Adaptive Data Analysis Using Information Theory , 2015, AISTATS.
[9] Matus Telgarsky,et al. Benefits of Depth in Neural Networks , 2016, COLT.
[10] Cynthia Dwork,et al. Calibrating Noise to Sensitivity in Private Data Analysis , 2006, TCC.
[11] Shai Ben-David,et al. Understanding Machine Learning: From Theory to Algorithms , 2014 .
[12] Emiliano De Cristofaro,et al. LOGAN: Membership Inference Attacks Against Generative Models , 2017, Proc. Priv. Enhancing Technol..
[13] Shai Ben-David,et al. On the difficulty of approximately maximizing agreements , 2000, J. Comput. Syst. Sci..
[14] E. Ordentlich,et al. Inequalities for the L1 Deviation of the Empirical Distribution , 2003 .
[15] Tengyu Ma,et al. On the Ability of Neural Nets to Express Distributions , 2017, COLT.
[16] Vitaly Shmatikov,et al. Overlearning Reveals Sensitive Attributes , 2019, ICLR.
[17] Andrew R. Barron,et al. Universal approximation bounds for superpositions of a sigmoidal function , 1993, IEEE Trans. Inf. Theory.
[18] Vitaly Shmatikov,et al. Auditing Data Provenance in Text-Generation Models , 2018, KDD.
[19] Ram Rajagopal,et al. Context-Aware Generative Adversarial Privacy , 2017, Entropy.
[20] Julia Rubin,et al. Fairness Definitions Explained , 2018, 2018 IEEE/ACM International Workshop on Software Fairness (FairWare).
[21] Reza Shokri,et al. Comprehensive Privacy Analysis of Deep Learning: Stand-alone and Federated Learning under Passive and Active White-box Inference Attacks , 2018, ArXiv.
[22] Amit Daniely,et al. Depth Separation for Neural Networks , 2017, COLT.
[23] James Zou,et al. How Much Does Your Data Exploration Overfit? Controlling Bias via Information Usage , 2015, IEEE Transactions on Information Theory.
[24] Vitaly Shmatikov,et al. Membership Inference Attacks Against Machine Learning Models , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[25] Cynthia Dwork,et al. Differential Privacy: A Survey of Results , 2008, TAMC.
[26] André Elisseeff,et al. Stability and Generalization , 2002, J. Mach. Learn. Res..
[27] Daniel Bernau,et al. Monte Carlo and Reconstruction Membership Inference Attacks against Generative Models , 2019, Proc. Priv. Enhancing Technol..
[28] David Berthelot,et al. High-Fidelity Extraction of Neural Network Models , 2019, ArXiv.
[29] Vitaly Shmatikov,et al. Machine Learning Models that Remember Too Much , 2017, CCS.
[30] Moni Naor,et al. Our Data, Ourselves: Privacy Via Distributed Noise Generation , 2006, EUROCRYPT.
[31] Mansoor I. Yousefi,et al. Bounds on the Approximation Power of Feedforward Neural Networks , 2018, ICML.
[32] Ohad Shamir,et al. The Power of Depth for Feedforward Neural Networks , 2015, COLT.
[33] Úlfar Erlingsson,et al. The Secret Sharer: Measuring Unintended Neural Network Memorization & Extracting Secrets , 2018, ArXiv.
[34] Samy Bengio,et al. Understanding deep learning requires rethinking generalization , 2016, ICLR.
[35] Ohad Shamir,et al. Learning and generalization with the information bottleneck , 2008, Theoretical Computer Science.
[36] Maxim Raginsky,et al. Information-theoretic analysis of generalization capability of learning algorithms , 2017, NIPS.
[37] Graeme Smith,et al. A Tight Uniform Continuity Bound for Equivocation , 2020, 2020 IEEE International Symposium on Information Theory (ISIT).