Counterfactually-Augmented SNLI Training Data Does Not Yield Better Generalization Than Unaugmented Data
暂无分享,去创建一个
[1] Samuel R. Bowman,et al. A Broad-Coverage Challenge Corpus for Sentence Understanding through Inference , 2017, NAACL.
[2] Eduard Hovy,et al. Learning the Difference that Makes a Difference with Counterfactually-Augmented Data , 2020, ICLR.
[3] Carolyn Penstein Rosé,et al. Stress Test Evaluation for Natural Language Inference , 2018, COLING.
[4] Christopher Potts,et al. A large annotated corpus for learning natural language inference , 2015, EMNLP.
[5] Omer Levy,et al. Annotation Artifacts in Natural Language Inference Data , 2018, NAACL.
[6] Omer Levy,et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.
[7] R. Thomas McCoy,et al. Right for the Wrong Reasons: Diagnosing Syntactic Heuristics in Natural Language Inference , 2019, ACL.
[8] Daniel Khashabi,et al. More Bang for Your Buck: Natural Perturbation for Robust Question Answering , 2020, EMNLP.
[9] Rachel Rudinger,et al. Hypothesis Only Baselines in Natural Language Inference , 2018, *SEMEVAL.
[10] Masatoshi Tsuchiya,et al. Performance Impact Caused by Hidden Bias of Training Data for Recognizing Textual Entailment , 2018, LREC.
[11] Mohit Bansal,et al. Adversarial NLI: A New Benchmark for Natural Language Understanding , 2020, ACL.
[12] Samuel R. Bowman,et al. Collecting Entailment Data for Pretraining: New Protocols and Negative Results , 2020, EMNLP.
[13] Omer Levy,et al. GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding , 2018, BlackboxNLP@EMNLP.
[14] Jan Gorodkin,et al. Comparing two K-category assignments by a K-category correlation coefficient , 2004, Comput. Biol. Chem..
[15] Marius Mosbach,et al. On the Stability of Fine-tuning BERT: Misconceptions, Explanations, and Strong Baselines , 2020, ArXiv.
[16] Alex Wang,et al. jiant: A Software Toolkit for Research on General-Purpose Text Understanding Models , 2020, ACL.