暂无分享,去创建一个
Matthias Bethge | Oliver Bringmann | Wieland Brendel | Peter Gehler | Evgenia Rusak | Steffen Schneider
[1] Matthias Bethge,et al. Improving robustness against common corruptions by covariate shift adaptation , 2020, NeurIPS.
[2] Aritra Ghosh,et al. Robust Loss Functions under Label Noise for Deep Neural Networks , 2017, AAAI.
[3] Zhuowen Tu,et al. Aggregated Residual Transformations for Deep Neural Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[4] D. Song,et al. The Many Faces of Robustness: A Critical Analysis of Out-of-Distribution Generalization , 2020, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).
[5] Sébastien Marcel,et al. Torchvision the machine-vision package of torch , 2010, ACM Multimedia.
[6] Samy Bengio,et al. Understanding deep learning requires rethinking generalization , 2016, ICLR.
[7] Boris Katz,et al. ObjectNet: A large-scale bias-controlled dataset for pushing the limits of object recognition models , 2019, NeurIPS.
[8] Ole Tange,et al. GNU Parallel: The Command-Line Power Tool , 2011, login Usenix Mag..
[9] Jae-Gil Lee,et al. Learning from Noisy Labels with Deep Neural Networks: A Survey , 2020, ArXiv.
[10] Geoffrey French,et al. Self-ensembling for visual domain adaptation , 2017, ICLR.
[11] Kaiming He,et al. Exploring the Limits of Weakly Supervised Pretraining , 2018, ECCV.
[12] Justin Gilmer,et al. MNIST-C: A Robustness Benchmark for Computer Vision , 2019, ArXiv.
[13] Stefano Ermon,et al. A DIRT-T Approach to Unsupervised Domain Adaptation , 2018, ICLR.
[14] James Bailey,et al. Symmetric Cross Entropy for Robust Learning With Noisy Labels , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[15] Matthias Bethge,et al. Increasing the robustness of DNNs against image corruptions by playing the Game of Noise , 2020, ArXiv.
[16] Geoffrey E. Hinton,et al. Distilling the Knowledge in a Neural Network , 2015, ArXiv.
[17] Trevor Darrell,et al. Fully Test-time Adaptation by Entropy Minimization , 2020, ArXiv.
[18] Dirk Merkel,et al. Docker: lightweight Linux containers for consistent development and deployment , 2014 .
[19] Yueming Lyu,et al. Curriculum Loss: Robust Learning and Generalization against Label Corruption , 2019, ICLR.
[20] Tom Schaul,et al. StarCraft II: A New Challenge for Reinforcement Learning , 2017, ArXiv.
[21] Yoshua Bengio,et al. Semi-supervised Learning by Entropy Minimization , 2004, CAP.
[22] Mert R. Sabuncu,et al. Generalized Cross Entropy Loss for Training Deep Neural Networks with Noisy Labels , 2018, NeurIPS.
[23] Alexander Gammerman,et al. Learning by Transduction , 1998, UAI.
[24] Thomas G. Dietterich,et al. Benchmarking Neural Network Robustness to Common Corruptions and Perturbations , 2018, ICLR.
[25] Balaji Lakshminarayanan,et al. AugMix: A Simple Data Processing Method to Improve Robustness and Uncertainty , 2020, ICLR.
[26] Quoc V. Le,et al. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks , 2019, ICML.
[27] Dong-Hyun Lee,et al. Pseudo-Label : The Simple and Efficient Semi-Supervised Learning Method for Deep Neural Networks , 2013 .
[28] François Chollet,et al. Xception: Deep Learning with Depthwise Separable Convolutions , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[29] Joel Nothman,et al. SciPy 1.0-Fundamental Algorithms for Scientific Computing in Python , 2019, ArXiv.
[30] Paul R. Cohen,et al. Empirical Comparison of "Hard" and "Soft" Label Propagation for Relational Classification , 2007, ILP.
[31] Nic Ford,et al. Adversarial Examples Are a Natural Consequence of Test Error in Noise , 2019, ICML.
[32] Jakub W. Pachocki,et al. Dota 2 with Large Scale Deep Reinforcement Learning , 2019, ArXiv.
[33] Matthias Bethge,et al. Generalisation in humans and deep neural networks , 2018, NeurIPS.
[34] Alexander S. Ecker,et al. Benchmarking Robustness in Object Detection: Autonomous Driving when Winter is Coming , 2019, ArXiv.
[35] Michał Grochowski,et al. Data augmentation for improving deep learning in image classification problem , 2018, 2018 International Interdisciplinary PhD Workshop (IIPhDW).
[36] Deyu Meng,et al. Learning Adaptive Loss for Robust Learning with Noisy Labels , 2020, ArXiv.
[37] Yang Chen,et al. Multi-Source Domain Adaptation and Semi-Supervised Domain Adaptation with Focus on Visual Domain Adaptation Challenge 2019 , 2019, ArXiv.
[38] Longin Jan Latecki,et al. Transductive Domain Adaptation with Affinity Learning , 2015, CIKM.
[39] Quoc V. Le,et al. Self-Training With Noisy Student Improves ImageNet Classification , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[40] Lina J. Karam,et al. A Study and Comparison of Human and Deep Learning Recognition Performance under Visual Distortions , 2017, 2017 26th International Conference on Computer Communication and Networks (ICCCN).
[41] Li Fei-Fei,et al. ImageNet: A large-scale hierarchical image database , 2009, CVPR.
[42] Silvio Savarese,et al. Learning Transferrable Representations for Unsupervised Domain Adaptation , 2016, NIPS.
[43] Michael S. Bernstein,et al. ImageNet Large Scale Visual Recognition Challenge , 2014, International Journal of Computer Vision.
[44] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[45] Yuan Yu,et al. TensorFlow: A system for large-scale machine learning , 2016, OSDI.
[46] Alexei A. Efros,et al. Test-Time Training for Out-of-Distribution Generalization , 2019, ArXiv.
[47] Yang Liu,et al. Learning with Instance-Dependent Label Noise: A Sample Sieve Approach , 2021, ICLR.
[48] Manfred K. Warmuth,et al. Robust Bi-Tempered Logistic Loss Based on Bregman Divergences , 2019, NeurIPS.
[49] Mark Chen,et al. Language Models are Few-Shot Learners , 2020, NeurIPS.
[50] Carsten Rother,et al. Benchmarking the Robustness of Semantic Segmentation Models with Respect to Common Corruptions , 2019, International Journal of Computer Vision.
[51] Matthias Bethge,et al. ImageNet-trained CNNs are biased towards texture; increasing shape bias improves accuracy and robustness , 2018, ICLR.
[52] Luca Antiga,et al. Automatic differentiation in PyTorch , 2017 .