Diagnosing Transformers in Task-Oriented Semantic Parsing

Modern task-oriented semantic parsing approaches typically use seq2seq transformers to map textual utterances to semantic frames comprised of intents and slots. While these models are empirically strong, their specific strengths and weaknesses have largely remained unexplored. In this work, we study BART (Lewis et al., 2020) and XLM-R (Conneau et al., 2020), two state-of-the-art parsers, across both monolingual and multilingual settings. Our experiments yield several key results: transformer-based parsers struggle not only with disambiguating intents/slots, but surprisingly also with producing syntacticallyvalid frames. Though pre-training imbues transformers with syntactic inductive biases, we find the ambiguity of copying utterance spans into frames often leads to tree invalidity, indicating span extraction is a major bottleneck for current parsers. However, as a silver lining, we show transformer-based parsers give sufficient indicators for whether a frame is likely to be correct or incorrect, making them easier to deploy in production settings.

[1]  Arash Einolghozati,et al.  Likelihood Ratios and Generative Classifiers for Unsupervised Out-of-Domain Detection In Task Oriented Dialog , 2019, AAAI.

[2]  Sonal Gupta,et al.  Semantic Parsing for Task Oriented Dialog using Hierarchical Representations , 2018, EMNLP.

[3]  Shrey Desai,et al.  Calibration of Pre-trained Transformers , 2020, EMNLP.

[4]  Alex Kulesza,et al.  Confidence Estimation for Machine Translation , 2004, COLING.

[5]  Benoît Sagot,et al.  What Does BERT Learn about the Structure of Language? , 2019, ACL.

[6]  Asish Ghoshal,et al.  Low-Resource Domain Adaptation for Compositional Task-Oriented Semantic Parsing , 2020, EMNLP.

[7]  Koby Crammer,et al.  Active Learning with Confidence , 2008, ACL.

[8]  Omer Levy,et al.  BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension , 2019, ACL.

[9]  Emine Yilmaz,et al.  Uncertainty and Traffic-Aware Active Learning for Semantic Parsing , 2020, INTEXSEMPAR.

[10]  Dipanjan Das,et al.  BERT Rediscovers the Classical NLP Pipeline , 2019, ACL.

[11]  Yoav Goldberg,et al.  Assessing BERT's Syntactic Abilities , 2019, ArXiv.

[12]  Haoran Li,et al.  MTOP: A Comprehensive Multilingual Task-Oriented Semantic Parsing Benchmark , 2020, EACL.

[13]  Veselin Stoyanov,et al.  Unsupervised Cross-lingual Representation Learning at Scale , 2019, ACL.

[14]  Dominique Estival,et al.  Active learning for deep semantic parsing , 2018, ACL.

[15]  Lukasz Kaiser,et al.  Attention is All you Need , 2017, NIPS.

[16]  Haoran Li,et al.  Conversational Semantic Parsing , 2020, EMNLP.