Statistically Robust Neural Network Classification
暂无分享,去创建一个
[1] Jason Weston,et al. Vicinal Risk Minimization , 2000, NIPS.
[2] J. I. Minnix. Fault tolerance of the backpropagation neural network trained on noisy inputs , 1992, [Proceedings 1992] IJCNN International Joint Conference on Neural Networks.
[3] Sida I. Wang,et al. Dropout Training as Adaptive Regularization , 2013, NIPS.
[4] Guozhong An,et al. The Effects of Adding Noise During Backpropagation Training on a Generalization Performance , 1996, Neural Computation.
[5] Rémi Munos,et al. Adaptive strategy for stratified Monte Carlo sampling , 2015, J. Mach. Learn. Res..
[6] Shai Ben-David,et al. Understanding Machine Learning: From Theory to Algorithms , 2014 .
[7] Luca Rigazio,et al. Towards Deep Neural Network Architectures Robust to Adversarial Examples , 2014, ICLR.
[8] Matus Telgarsky,et al. Spectrally-normalized margin bounds for neural networks , 2017, NIPS.
[9] John C. Duchi,et al. Learning Models with Uniform Performance via Distributionally Robust Optimization , 2018, ArXiv.
[10] Matthias Bethge,et al. Towards the first adversarially robust neural network model on MNIST , 2018, ICLR.
[11] Aleksander Madry,et al. Adversarially Robust Generalization Requires More Data , 2018, NeurIPS.
[12] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[13] Quoc V. Le,et al. Adding Gradient Noise Improves Learning for Very Deep Networks , 2015, ArXiv.
[14] Ananthram Swami,et al. The Limitations of Deep Learning in Adversarial Settings , 2015, 2016 IEEE European Symposium on Security and Privacy (EuroS&P).
[15] N. Hengartner,et al. Simulation and Estimation of Extreme Quantiles and Extreme Probabilities , 2011 .
[16] Nic Ford,et al. Adversarial Examples Are a Natural Consequence of Test Error in Noise , 2019, ICML.
[17] Swarat Chaudhuri,et al. AI2: Safety and Robustness Certification of Neural Networks with Abstract Interpretation , 2018, 2018 IEEE Symposium on Security and Privacy (SP).
[18] D Zipser,et al. Learning the hidden structure of speech. , 1988, The Journal of the Acoustical Society of America.
[19] Alexander J. Smola,et al. Detecting and Correcting for Label Shift with Black Box Predictors , 2018, ICML.
[20] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[21] Ameet Talwalkar,et al. Foundations of Machine Learning , 2012, Adaptive computation and machine learning.
[22] C. Lee Giles,et al. An analysis of noise in recurrent neural networks: convergence and generalization , 1996, IEEE Trans. Neural Networks.
[23] Mehryar Mohri,et al. Rademacher Complexity Bounds for Non-I.I.D. Processes , 2008, NIPS.
[24] Seyed-Mohsen Moosavi-Dezfooli,et al. DeepFool: A Simple and Accurate Method to Fool Deep Neural Networks , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[25] Valentina Zantedeschi,et al. Efficient Defenses Against Adversarial Attacks , 2017, AISec@CCS.
[26] Jocelyn Sietsma,et al. Creating artificial neural networks that generalize , 1991, Neural Networks.
[27] Kannan Ramchandran,et al. Rademacher Complexity for Adversarially Robust Generalization , 2018, ICML.
[28] J. Zico Kolter,et al. Provable defenses against adversarial examples via the convex outer adversarial polytope , 2017, ICML.
[29] J. Zico Kolter,et al. Certified Adversarial Robustness via Randomized Smoothing , 2019, ICML.
[30] Yee Whye Teh,et al. A Statistical Approach to Assessing Neural Network Robustness , 2018, ICLR.
[31] Christopher M. Bishop,et al. Current address: Microsoft Research, , 2022 .
[32] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[33] Mark S. Squillante,et al. PROVEN: Certifying Robustness of Neural Networks with a Probabilistic Approach , 2018, ICML.
[34] Saeed Mahloujifar,et al. Adversarial Risk and Robustness: General Definitions and Implications for the Uniform Distribution , 2018, NeurIPS.
[35] Petri Koistinen,et al. Using additive noise in back-propagation training , 1992, IEEE Trans. Neural Networks.
[36] Geoffrey E. Hinton,et al. Speech recognition with deep recurrent neural networks , 2013, 2013 IEEE International Conference on Acoustics, Speech and Signal Processing.
[37] Thomas G. Dietterich,et al. Benchmarking Neural Network Robustness to Common Corruptions and Perturbations , 2018, ICLR.
[38] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[39] Dan Boneh,et al. Ensemble Adversarial Training: Attacks and Defenses , 2017, ICLR.
[40] Pascal Frossard,et al. Analysis of classifiers’ robustness to adversarial perturbations , 2015, Machine Learning.
[41] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[42] Neil D. Lawrence,et al. Dataset Shift in Machine Learning , 2009 .
[43] Geoffrey E. Hinton,et al. Experiments on Learning by Back Propagation. , 1986 .
[44] Junfeng Yang,et al. Efficient Formal Safety Analysis of Neural Networks , 2018, NeurIPS.
[45] Aditi Raghunathan,et al. Certified Defenses against Adversarial Examples , 2018, ICLR.
[46] Andrew R. Webb,et al. Functional approximation by feed-forward networks: a least-squares approach to generalization , 1994, IEEE Trans. Neural Networks.
[47] Nikos Komodakis,et al. Wide Residual Networks , 2016, BMVC.
[48] Sergey Ioffe,et al. Rethinking the Inception Architecture for Computer Vision , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[49] M. Talagrand,et al. Probability in Banach Spaces: Isoperimetry and Processes , 1991 .