A Hybrid Neural Network RBERT-C Based on Pre-trained RoBERTa and CNN for User Intent Classification

User intent classification plays a critical role in identifying the interests of users in question-answering and spoken dialog systems. The question texts of these systems are usually short and their conveyed semantic information are frequently insufficient. Therefore, the accuracy of user intent classification related to user satisfaction may be affected. To address the problem, this paper proposes a hybrid neural network named RBERT-C for text classification to capture user intent. The network uses the Chinese pre-trained RoBERTa to initialize representation layer parameters. Then, it obtains question representations through a bidirectional transformer structure and extracts essential features using a Convolutional Neural Network after question representation modeling. The evaluation is based on the publicly available dataset ECDT containing 3736 labeled sentences. Experimental result indicates that our model RBERT-C achieves a F1 score of 0.96 and an accuracy of 0.96, outperforming a number of baseline methods.

[1]  James R. Glass,et al.  Query understanding enhanced by hierarchical parsing structures , 2013, 2013 IEEE Workshop on Automatic Speech Recognition and Understanding.

[2]  Tong Zhang,et al.  Deep Pyramid Convolutional Neural Networks for Text Categorization , 2017, ACL.

[3]  Tianyong Hao,et al.  A Feature Extraction and Expansion-Based Approach for Question Target Identification and Classification , 2017, CCIR.

[4]  Yoon Kim,et al.  Convolutional Neural Networks for Sentence Classification , 2014, EMNLP.

[5]  Francesco Caltagirone,et al.  Snips Voice Platform: an embedded Spoken Language Understanding system for private-by-design voice interfaces , 2018, ArXiv.

[6]  Qingyao Wu,et al.  Leveraging question target word features through semantic relation expansion for answer type classification , 2017, Knowl. Based Syst..

[7]  Jin Liu,et al.  Attention-based BiGRU-CNN for Chinese question classification , 2019, Journal of Ambient Intelligence and Humanized Computing.

[8]  Jian Zhang,et al.  Using Convolutional Neural Network with BERT for Intent Determination , 2019, 2019 International Conference on Asian Language Processing (IALP).

[9]  Gökhan Tür,et al.  What is left to be understood in ATIS? , 2010, 2010 IEEE Spoken Language Technology Workshop.

[10]  Bing Liu,et al.  Attention-Based Recurrent Neural Network Models for Joint Intent Detection and Slot Filling , 2016, INTERSPEECH.

[11]  Geoffrey E. Hinton,et al.  Learning distributed representations of concepts. , 1989 .

[12]  Geoffrey Zweig,et al.  Joint semantic utterance classification and slot filling with recursive neural networks , 2014, 2014 IEEE Spoken Language Technology Workshop (SLT).

[13]  Ming-Wei Chang,et al.  BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.

[14]  Bing Liu,et al.  Joint Online Spoken Language Understanding and Language Modeling With Recurrent Neural Networks , 2016, SIGDIAL Conference.

[15]  Ruhi Sarikaya,et al.  Convolutional neural network based triangular CRF for joint intent detection and slot filling , 2013, 2013 IEEE Workshop on Automatic Speech Recognition and Understanding.

[16]  Tianyong Hao,et al.  Capsule-Based Bidirectional Gated Recurrent Unit Networks for Question Target Classification , 2018, CCIR.

[17]  Tianyong Hao,et al.  A Feature-Enriched Method for User Intent Classification by Leveraging Semantic Tag Expansion , 2018, NLPCC.

[18]  Benoît Sagot,et al.  What Does BERT Learn about the Structure of Language? , 2019, ACL.

[19]  Chengyang Zhang,et al.  Sentiment-Aware Short Text Classification Based on Convolutional Neural Network and Attention , 2019, 2019 IEEE 31st International Conference on Tools with Artificial Intelligence (ICTAI).

[20]  Andreas Stolcke,et al.  Recurrent neural network and LSTM models for lexical utterance classification , 2015, INTERSPEECH.

[21]  Tianyong Hao,et al.  A WordNet Expansion-Based Approach for Question Targets Identification and Classification , 2015, CCL.

[22]  George Kurian,et al.  Google's Neural Machine Translation System: Bridging the Gap between Human and Machine Translation , 2016, ArXiv.

[23]  Talaat Khalil,et al.  Cross-lingual intent classification in a low resource industrial setting , 2019, EMNLP.