暂无分享,去创建一个
[1] Andrey Malinin,et al. Reverse KL-Divergence Training of Prior Networks: Improved Uncertainty and Adversarial Robustness , 2019, NeurIPS.
[2] Francis M. Tyers,et al. Common Voice: A Massively-Multilingual Speech Corpus , 2020, LREC.
[3] Thomas G. Dietterich,et al. Benchmarking Neural Network Robustness to Common Corruptions and Perturbations , 2018, ICLR.
[4] Matthew D. Zeiler. ADADELTA: An Adaptive Learning Rate Method , 2012, ArXiv.
[5] D. Song,et al. The Many Faces of Robustness: A Critical Analysis of Out-of-Distribution Generalization , 2020, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).
[6] Geoffrey E. Hinton,et al. Distilling the Knowledge in a Neural Network , 2015, ArXiv.
[7] Charles Blundell,et al. Simple and Scalable Predictive Uncertainty Estimation using Deep Ensembles , 2016, NIPS.
[8] Myle Ott,et al. Scaling Neural Machine Translation , 2018, WMT.
[9] John Schulman,et al. Concrete Problems in AI Safety , 2016, ArXiv.
[10] Graham Neubig,et al. Understanding Knowledge Distillation in Non-autoregressive Machine Translation , 2020, ICLR.
[11] Jean Carletta,et al. The AMI meeting corpus , 2005 .
[12] Dmitry Vetrov,et al. Pitfalls of In-Domain Uncertainty Estimation and Ensembling in Deep Learning , 2020, ICLR.
[13] Rico Sennrich,et al. Neural Machine Translation of Rare Words with Subword Units , 2015, ACL.
[14] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[15] Dawn Song,et al. Natural Adversarial Examples , 2019, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[16] Kaiming He,et al. Accurate, Large Minibatch SGD: Training ImageNet in 1 Hour , 2017, ArXiv.
[17] Yarin Gal,et al. BatchBALD: Efficient and Diverse Batch Acquisition for Deep Bayesian Active Learning , 2019, NeurIPS.
[18] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[19] Mark J. F. Gales,et al. Predictive Uncertainty Estimation via Prior Networks , 2018, NeurIPS.
[20] M. Gales,et al. Uncertainty in Structured Prediction , 2020, ArXiv.
[21] Sanjeev Khudanpur,et al. Librispeech: An ASR corpus based on public domain audio books , 2015, 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[22] Andrey Malinin,et al. Uncertainty estimation in deep learning with application to spoken language assessment , 2019 .
[23] Alexandr A. Kalinin,et al. Albumentations: fast and flexible image augmentations , 2018, Inf..
[24] Alexander M. Rush,et al. Sequence-Level Knowledge Distillation , 2016, EMNLP.
[25] Luke S. Zettlemoyer,et al. Transformers with convolutional context for ASR , 2019, ArXiv.
[26] Jasper Snoek,et al. Training independent subnetworks for robust prediction , 2020, ICLR.
[27] Sebastian Nowozin,et al. Hydra: Preserving Ensemble Diversity for Model Distillation , 2020, ArXiv.
[28] Sebastian Nowozin,et al. Can You Trust Your Model's Uncertainty? Evaluating Predictive Uncertainty Under Dataset Shift , 2019, NeurIPS.
[29] Salim Roukos,et al. Bleu: a Method for Automatic Evaluation of Machine Translation , 2002, ACL.
[30] Matthijs Douze,et al. Fixing the train-test resolution discrepancy , 2019, NeurIPS.
[31] Garrison W. Cottrell,et al. ReZero is All You Need: Fast Convergence at Large Depth , 2020, UAI.
[32] T. Minka. Estimating a Dirichlet distribution , 2012 .
[33] David Wagner,et al. Adversarial Examples Are Not Easily Detected: Bypassing Ten Detection Methods , 2017, AISec@CCS.
[34] Mark J. F. Gales,et al. Ensemble Approaches for Uncertainty in Spoken Language Assessment , 2020, INTERSPEECH.
[35] Andrey Malinin,et al. Ensemble Distribution Distillation , 2019, ICLR.
[36] Li Fei-Fei,et al. ImageNet: A large-scale hierarchical image database , 2009, CVPR.
[37] Yarin Gal,et al. Understanding Measures of Uncertainty for Adversarial Example Detection , 2018, UAI.
[38] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[39] Matt Post,et al. A Call for Clarity in Reporting BLEU Scores , 2018, WMT.
[40] Zoubin Ghahramani,et al. Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning , 2015, ICML.