暂无分享,去创建一个
[1] Guillaume Lample,et al. Cross-lingual Language Model Pretraining , 2019, NeurIPS.
[2] Boris Ginsburg,et al. Correction of Automatic Speech Recognition with Transformer Sequence-To-Sequence Model , 2019, ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[3] Alexander I. Rudnicky,et al. Expanding the Scope of the ATIS Task: The ATIS-3 Corpus , 1994, HLT.
[4] Wilson L. Taylor,et al. “Cloze Procedure”: A New Tool for Measuring Readability , 1953 .
[5] Mitchell P. Marcus,et al. Text Chunking using Transformation-Based Learning , 1995, VLC@ACL.
[6] Omer Levy,et al. BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension , 2019, ACL.
[7] Yun-Nung (Vivian) Chen,et al. Learning Asr-Robust Contextualized Embeddings for Spoken Language Understanding , 2019, ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[8] George R. Doddington,et al. The ATIS Spoken Language Systems Pilot Corpus , 1990, HLT.
[9] Gokhan Tur,et al. Joint Contextual Modeling for ASR Correction and Language Understanding , 2020, ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[10] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[11] Philip S. Yu,et al. Joint Slot Filling and Intent Detection via Capsule Neural Networks , 2018, ACL.
[12] Ilya Sutskever,et al. Language Models are Unsupervised Multitask Learners , 2019 .
[13] Gökhan Tür,et al. What is left to be understood in ATIS? , 2010, 2010 IEEE Spoken Language Technology Workshop.
[14] Erik F. Tjong Kim Sang,et al. Introduction to the CoNLL-2003 Shared Task: Language-Independent Named Entity Recognition , 2003, CoNLL.
[15] Wen Wang,et al. BERT for Joint Intent Classification and Slot Filling , 2019, ArXiv.
[16] Quoc V. Le,et al. ELECTRA: Pre-training Text Encoders as Discriminators Rather Than Generators , 2020, ICLR.
[17] Maryam Fazel-Zarandi,et al. Data Augmentation for Training Dialog Models Robust to Speech Recognition Errors , 2020, NLP4CONVAI.
[18] Hermann Ney,et al. Lattice decoding and rescoring with long-Span neural network language models , 2014, INTERSPEECH.
[19] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.
[20] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[21] Tie-Yan Liu,et al. MC-BERT: Efficient Language Pre-Training via a Meta Controller , 2020, ArXiv.
[22] Hermann Ney,et al. Language Modeling with Deep Transformers , 2019, INTERSPEECH.