暂无分享,去创建一个
[1] Honglak Lee,et al. An Analysis of Single-Layer Networks in Unsupervised Feature Learning , 2011, AISTATS.
[2] Samy Bengio,et al. Adversarial examples in the physical world , 2016, ICLR.
[3] J. Zico Kolter,et al. Learning perturbation sets for robust machine learning , 2020, ICLR.
[4] Stephan J. Garbin,et al. Harmonic Networks: Deep Translation and Rotation Equivariance , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[5] Benjamin Recht,et al. Do CIFAR-10 Classifiers Generalize to CIFAR-10? , 2018, ArXiv.
[6] Ananthram Swami,et al. The Limitations of Deep Learning in Adversarial Settings , 2015, 2016 IEEE European Symposium on Security and Privacy (EuroS&P).
[7] Andrew Zisserman,et al. Spatial Transformer Networks , 2015, NIPS.
[8] George J. Pappas,et al. Model-Based Robust Deep Learning , 2020, ArXiv.
[9] Max Welling,et al. Group Equivariant Convolutional Networks , 2016, ICML.
[10] Geoffrey E. Hinton. A Parallel Computation that Assigns Canonical Object-Based Frames of Reference , 1981, IJCAI.
[11] Hang Su,et al. Benchmarking Adversarial Robustness on Image Classification , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[12] Ekin D. Cubuk,et al. A Fourier Perspective on Model Robustness in Computer Vision , 2019, NeurIPS.
[13] Kouichi Sakurai,et al. One Pixel Attack for Fooling Deep Neural Networks , 2017, IEEE Transactions on Evolutionary Computation.
[14] Lina J. Karam,et al. Quality Resilient Deep Neural Networks , 2017, ArXiv.
[15] Thomas G. Dietterich,et al. Benchmarking Neural Network Robustness to Common Corruptions and Perturbations , 2018, ICLR.
[16] Arnold W. M. Smeulders,et al. Structured Receptive Fields in CNNs , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[17] Yair Weiss,et al. Why do deep convolutional networks generalize so poorly to small image transformations? , 2018, J. Mach. Learn. Res..
[18] Joachim M. Buhmann,et al. TI-POOLING: Transformation-Invariant Pooling for Feature Learning in Convolutional Neural Networks , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[19] D. Song,et al. The Many Faces of Robustness: A Critical Analysis of Out-of-Distribution Generalization , 2020, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).
[20] Ivan Sosnovik,et al. DISCO: accurate Discrete Scale Convolutions , 2021, BMVC.
[21] In So Kweon,et al. Revisiting Batch Normalization for Improving Corruption Robustness , 2021, 2021 IEEE Winter Conference on Applications of Computer Vision (WACV).
[22] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[23] Matthias Bethge,et al. Increasing the robustness of DNNs against image corruptions by playing the Game of Noise , 2020, ArXiv.
[24] Matthias Bethge,et al. Comparing deep neural networks against humans: object recognition when the signal gets weaker , 2017, ArXiv.
[25] Matthias Bethge,et al. Improving robustness against common corruptions by covariate shift adaptation , 2020, NeurIPS.
[26] Yi Li,et al. Deformable Convolutional Networks , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[27] Stéphane Mallat,et al. Invariant Scattering Convolution Networks , 2012, IEEE transactions on pattern analysis and machine intelligence.
[28] Yann LeCun,et al. Transformation Invariance in Pattern Recognition-Tangent Distance and Tangent Propagation , 1996, Neural Networks: Tricks of the Trade.
[29] Seyed-Mohsen Moosavi-Dezfooli,et al. Geometric Robustness of Deep Networks: Analysis and Improvement , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[30] Daniel E. Worrall,et al. Deep Scale-spaces: Equivariance Over Scale , 2019, NeurIPS.
[31] Dimitris N. Metaxas,et al. SelfNorm and CrossNorm for Out-of-Distribution Robustness , 2021, ArXiv.
[32] Aleksander Madry,et al. Exploring the Landscape of Spatial Robustness , 2017, ICML.
[33] Inderjit S. Dhillon,et al. The Limitations of Adversarial Training and the Blind-Spot Attack , 2019, ICLR.
[34] Maurice Weiler,et al. Learning Steerable Filters for Rotation Equivariant CNNs , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[35] Ivan Sosnovik,et al. Scale Equivariance Improves Siamese Tracking , 2020, 2021 IEEE Winter Conference on Applications of Computer Vision (WACV).
[36] Matthias Bethge,et al. A Simple Way to Make Neural Networks Robust Against Diverse Image Corruptions , 2020, ECCV.
[37] Lina J. Karam,et al. A Study and Comparison of Human and Deep Learning Recognition Performance under Visual Distortions , 2017, 2017 26th International Conference on Computer Communication and Networks (ICCCN).
[38] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[39] David A. McAllester,et al. Object Detection with Discriminatively Trained Part Based Models , 2010, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[40] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[41] Seyed-Mohsen Moosavi-Dezfooli,et al. DeepFool: A Simple and Accurate Method to Fool Deep Neural Networks , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[42] Pascal Frossard,et al. Manitest: Are classifiers really invariant? , 2015, BMVC.
[43] Ivan Sosnovik,et al. Scale-Equivariant Steerable Networks , 2020, ICLR.
[44] David W. Jacobs,et al. Locally Scale-Invariant Convolutional Neural Networks , 2014, ArXiv.
[45] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[46] Arnold Smeulders,et al. Natural Perturbed Training for General Robustness of Neural Network Classifiers , 2021, ArXiv.