Rademacher Complexity for Adversarially Robust Generalization
暂无分享,去创建一个
[1] M. Talagrand,et al. Probability in Banach Spaces: Isoperimetry and Processes , 1991 .
[2] Peter L. Bartlett,et al. Efficient agnostic learning of neural networks with bounded fan-in , 1996, IEEE Trans. Inf. Theory.
[3] Peter L. Bartlett,et al. The Sample Complexity of Pattern Classification with Neural Networks: The Size of the Weights is More Important than the Size of the Network , 1998, IEEE Trans. Inf. Theory.
[4] Yoshua Bengio,et al. Gradient-based learning applied to document recognition , 1998, Proc. IEEE.
[5] Peter L. Bartlett,et al. Neural Network Learning - Theoretical Foundations , 1999 .
[6] Peter L. Bartlett,et al. Rademacher and Gaussian Complexities: Risk Bounds and Structural Results , 2003, J. Mach. Learn. Res..
[7] Simon Haykin,et al. GradientBased Learning Applied to Document Recognition , 2001 .
[8] V. Koltchinskii. Local Rademacher complexities and oracle inequalities in risk minimization , 2006, 0708.0083.
[9] Shie Mannor,et al. Robustness and Regularization of Support Vector Machines , 2008, J. Mach. Learn. Res..
[10] Shie Mannor,et al. Robust Regression and Lasso , 2008, IEEE Transactions on Information Theory.
[11] Shie Mannor,et al. Robustness and generalization , 2010, Machine Learning.
[12] Ameet Talwalkar,et al. Foundations of Machine Learning , 2012, Adaptive computation and machine learning.
[13] Geoffrey E. Hinton,et al. Speech recognition with deep recurrent neural networks , 2013, 2013 IEEE International Conference on Acoustics, Speech and Signal Processing.
[14] Jean-Philippe Vial,et al. Robust Optimization , 2021, ICORES.
[15] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[16] Luca Rigazio,et al. Towards Deep Neural Network Architectures Robust to Adversarial Examples , 2014, ICLR.
[17] M. Mohri,et al. Rademacher Complexity Margin Bounds for Learning with a Large Number of Classes , 2015 .
[18] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[19] Uri Shaham,et al. Understanding Adversarial Training: Increasing Local Stability of Neural Nets through Robust Optimization , 2015, ArXiv.
[20] Dale Schuurmans,et al. Learning with a Strong Adversary , 2015, ArXiv.
[21] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[22] Yuchen Zhang,et al. L1-regularized Neural Networks are Improperly Learnable in Polynomial Time , 2015, ICML.
[23] Yury Maximov,et al. Tight risk bounds for multi-class margin classifiers , 2016, Pattern Recognition and Image Analysis.
[24] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[25] Yuan Yu,et al. TensorFlow: A system for large-scale machine learning , 2016, OSDI.
[26] Seyed-Mohsen Moosavi-Dezfooli,et al. Robustness of classifiers: from adversarial to random noise , 2016, NIPS.
[27] Demis Hassabis,et al. Mastering the game of Go with deep neural networks and tree search , 2016, Nature.
[28] David A. Wagner,et al. Defensive Distillation is Not Robust to Adversarial Examples , 2016, ArXiv.
[29] Samy Bengio,et al. Understanding deep learning requires rethinking generalization , 2016, ICLR.
[30] Matus Telgarsky,et al. Spectrally-normalized margin bounds for neural networks , 2017, NIPS.
[31] David Wagner,et al. Adversarial Examples Are Not Easily Detected: Bypassing Ten Detection Methods , 2017, AISec@CCS.
[32] Aleksander Madry,et al. A Rotation and a Translation Suffice: Fooling CNNs with Simple Transformations , 2017, ArXiv.
[33] Ryan P. Adams,et al. Motivating the Rules of the Game for Adversarial Example Research , 2018, ArXiv.
[34] Elvis Dohmatob,et al. Limitations of adversarial robustness: strong No Free Lunch Theorem , 2018, ArXiv.
[35] J. Zico Kolter,et al. Scaling provable adversarial defenses , 2018, NeurIPS.
[36] Po-Ling Loh,et al. Adversarial Risk Bounds for Binary Classification via Function Transformation , 2018, ArXiv.
[37] Michael P. Wellman,et al. SoK: Security and Privacy in Machine Learning , 2018, 2018 IEEE European Symposium on Security and Privacy (EuroS&P).
[38] Somesh Jha,et al. Analyzing the Robustness of Nearest Neighbors to Adversarial Examples , 2017, ICML.
[39] Ohad Shamir,et al. Size-Independent Sample Complexity of Neural Networks , 2017, COLT.
[40] J. Zico Kolter,et al. Provable defenses against adversarial examples via the convex outer adversarial polytope , 2017, ICML.
[41] David A. McAllester,et al. A PAC-Bayesian Approach to Spectrally-Normalized Margin Bounds for Neural Networks , 2017, ICLR.
[42] David A. Wagner,et al. Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples , 2018, ICML.
[43] David A. Wagner,et al. Audio Adversarial Examples: Targeted Attacks on Speech-to-Text , 2018, 2018 IEEE Security and Privacy Workshops (SPW).
[44] Hamza Fawzi,et al. Adversarial vulnerability for any classifier , 2018, NeurIPS.
[45] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[46] Aditi Raghunathan,et al. Semidefinite relaxations for certifying robustness to adversarial examples , 2018, NeurIPS.
[47] Aleksander Madry,et al. Adversarially Robust Generalization Requires More Data , 2018, NeurIPS.
[48] John C. Duchi,et al. Certifying Some Distributional Robustness with Principled Adversarial Training , 2017, ICLR.
[49] Dawn Xiaodong Song,et al. Adversarial Examples for Generative Models , 2017, 2018 IEEE Security and Privacy Workshops (SPW).
[50] Aditi Raghunathan,et al. Certified Defenses against Adversarial Examples , 2018, ICLR.
[51] Prateek Mittal,et al. PAC-learning in the presence of evasion adversaries , 2018, NIPS 2018.
[52] Yi Zhang,et al. Stronger generalization bounds for deep nets via a compression approach , 2018, ICML.
[53] Pradeep Ravikumar,et al. On Adversarial Risk and Training , 2018, ArXiv.
[54] Andrea Montanari,et al. A mean field view of the landscape of two-layer neural networks , 2018, Proceedings of the National Academy of Sciences.
[55] Aleksander Madry,et al. Exploring the Landscape of Spatial Robustness , 2017, ICML.
[56] Elvis Dohmatob,et al. Generalized No Free Lunch Theorem for Adversarial Robustness , 2018, ICML.
[57] Aleksander Madry,et al. Robustness May Be at Odds with Accuracy , 2018, ICLR.
[58] David Tse,et al. Generalizable Adversarial Training via Spectral Normalization , 2018, ICLR.
[59] Yishay Mansour,et al. Improved generalization bounds for robust learning , 2018, ALT.
[60] Saeed Mahloujifar,et al. The Curse of Concentration in Robust Learning: Evasion and Poisoning Attacks from Concentration of Measure , 2018, AAAI.
[61] Ilya P. Razenshteyn,et al. Adversarial examples from computational constraints , 2018, ICML.