暂无分享,去创建一个
[1] Milan Straka,et al. CzeSL Grammatical Error Correction Dataset (CzeSL-GEC) , 2017 .
[2] Zhongjun He,et al. Robust Neural Machine Translation with Joint Textual and Phonetic Embedding , 2018, ACL.
[3] Lin Yang,et al. Squared English Word: A Method of Generating Glyph to Use Super Characters for Sentiment Analysis , 2019, AffCon@AAAI.
[4] Myle Ott,et al. fairseq: A Fast, Extensible Toolkit for Sequence Modeling , 2019, NAACL.
[5] Soroush Vosoughi,et al. Tweet2Vec: Learning Tweet Embeddings Using Character-level CNN-LSTM Encoder-Decoder , 2016, SIGIR.
[6] Eric P. Xing,et al. High-Frequency Component Helps Explain the Generalization of Convolutional Neural Networks , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[7] Yonatan Belinkov,et al. Synthetic and Natural Noise Both Break Neural Machine Translation , 2017, ICLR.
[8] Jürgen Schmidhuber,et al. Training Very Deep Networks , 2015, NIPS.
[9] Marcin Junczys-Dowmunt,et al. The University of Edinburgh’s systems submission to the MT task at IWSLT , 2018, IWSLT.
[10] Tao Meng,et al. On the Robustness of Language Encoders against Grammatical Errors , 2020, ACL.
[11] Aditi Raghunathan,et al. Robust Encodings: A Framework for Combating Adversarial Typos , 2020, ACL.
[12] Jun Suzuki,et al. Effective Adversarial Regularization for Neural Machine Translation , 2019, ACL.
[13] Frederick Liu,et al. Learning Character-level Compositionality with Visual Features , 2017, ACL.
[14] Felix Gräßer,et al. Aspect-Based Sentiment Analysis of Drug Reviews Applying Cross-Domain and Cross-Data Learning , 2018, DH.
[15] Graham Neubig,et al. Improving Robustness of Neural Machine Translation with Multi-task Learning , 2019, WMT.
[16] Nada Lavrac,et al. tax2vec: Constructing Interpretable Features from Taxonomies for Short Text Classification , 2019, Comput. Speech Lang..
[17] Torsten Zesch,et al. Measuring Contextual Fitness Using Error Contexts Extracted from the Wikipedia Revision History , 2012, EACL.
[18] Walt Detmar Meurers,et al. MERLIN : An Online Trilingual Learner Corpus Empirically Grounding the European Reference Levels in Authentic Learner Data , 2013 .
[19] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[20] Wei Wu,et al. Glyce: Glyph-vectors for Chinese Character Representations , 2019, NeurIPS.
[21] Jun Zhou,et al. cw2vec: Learning Chinese Word Embeddings with Stroke n-gram Information , 2018, AAAI.
[22] Dejing Dou,et al. HotFlip: White-Box Adversarial Examples for NLP , 2017, ArXiv.
[23] Chao Liu,et al. Radical Embedding: Delving Deeper to Chinese Radicals , 2015, ACL.
[24] Falcon Z. Dai,et al. Glyph-aware Embedding of Chinese Characters , 2017, SWCN@EMNLP.
[25] Christopher D. Manning,et al. Stanford Neural Machine Translation Systems for Spoken Language Domains , 2015, IWSLT.
[26] Rico Sennrich,et al. Has Machine Translation Achieved Human Parity? A Case for Document-level Evaluation , 2018, EMNLP.
[27] David Chiang,et al. Neural Machine Translation of Text from Non-Native Speakers , 2018, NAACL.
[28] Eric P. Xing,et al. What If We Simply Swap the Two Text Fragments? A Straightforward yet Effective Way to Test the Robustness of Methods to Confounding Signals in Nature Language Inference Tasks , 2018, AAAI.
[29] Alexander M. Rush,et al. Character-Aware Neural Language Models , 2015, AAAI.
[30] Bhuwan Dhingra,et al. Combating Adversarial Misspellings with Robust Word Recognition , 2019, ACL.
[31] Graham Neubig,et al. Improving Robustness of Machine Translation with Synthetic Noise , 2019, NAACL.
[32] Yong Cheng,et al. Robust Neural Machine Translation with Doubly Adversarial Inputs , 2019, ACL.
[33] Po-Sen Huang,et al. Achieving Verified Robustness to Symbol Substitutions via Interval Bound Propagation , 2019, EMNLP/IJCNLP.
[34] Guillaume Wisniewski,et al. Mining Naturally-occurring Corrections and Paraphrases from Wikipedia’s Revision History , 2022, LREC.
[35] Lijun Wu,et al. Achieving Human Parity on Automatic Chinese to English News Translation , 2018, ArXiv.
[36] Aditi Raghunathan,et al. Certified Robustness to Adversarial Word Substitutions , 2019, EMNLP.
[37] George Kurian,et al. Google's Neural Machine Translation System: Bridging the Gap between Human and Machine Translation , 2016, ArXiv.
[38] Xiang Zhang,et al. Which Encoding is the Best for Text Classification in Chinese, English, Japanese and Korean? , 2017, ArXiv.
[39] Hung-yi Lee,et al. Learning Chinese Word Representations From Glyphs Of Characters , 2017, EMNLP.
[40] Omer Levy,et al. Training on Synthetic Noise Improves Robustness to Natural Noise in Machine Translation , 2019, EMNLP.
[41] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[42] Xirong Li,et al. Deep Text Classification Can be Fooled , 2017, IJCAI.
[43] Dejing Dou,et al. HotFlip: White-Box Adversarial Examples for Text Classification , 2017, ACL.
[44] Kevin Duh,et al. Robsut Wrod Reocginiton via Semi-Character Recurrent Neural Network , 2016, AAAI.
[45] Eric P. Xing,et al. Select-additive learning: Improving generalization in multimodal sentiment analysis , 2016, 2017 IEEE International Conference on Multimedia and Expo (ICME).
[46] Yash Goyal,et al. Making the V in VQA Matter: Elevating the Role of Image Understanding in Visual Question Answering , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[47] Lin Yang,et al. Super Characters: A Conversion from Sentiment Classification to Image Classification , 2018, WASSA@EMNLP.
[48] Zachary C. Lipton,et al. How Much Reading Does Reading Comprehension Require? A Critical Investigation of Popular Benchmarks , 2018, EMNLP.
[49] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.