Detection and Interpretation of Opinion Expressions in Spoken Surveys

This paper describes a system for automatic opinion analysis from spoken messages collected in the context of a user satisfaction survey. Opinion analysis is performed from the perspective of opinion monitoring. A process is outlined for detecting segments expressing opinions in a speech signal. Methods are proposed for accepting or rejecting segments from messages that are not reliably analyzed due to the limitations of automatic speech recognition processes, for assigning opinion hypotheses to segments and for evaluating hypothesis opinion proportions. Specific language models are introduced for representing opinion concepts. These models are used for hypothesizing opinion carrying segments in a spoken message. Each segment is interpreted by a classifier based on the Adaboost algorithm which associates a pair of topic and polarity labels to each segment. The different processes are trained and evaluated on a telephone corpus collected in a deployed customer care service. The use of conditional random fields (CRFs) is also considered for detecting segments and results are compared for different types of data and approaches. By optimizing the choice of the strategy parameters, it is possible to estimate user opinion proportions with a Kullback-Leibler divergence of 0.047 bits with respect to the true proportions obtained with a manual annotation of the spoken messages. The proportions estimated with such a low divergence are accurate enough for monitoring user satisfaction over time.

[1]  Beth Levin,et al.  English Verb Classes and Alternations: A Preliminary Investigation , 1993 .

[2]  Janyce Wiebe,et al.  Tracking Point of View in Narrative , 1994, Comput. Linguistics.

[3]  Dirk Noël Beth Levin. English Verb Classes and Alternations: A Preliminary Investigation , 1995 .

[4]  Vasileios Hatzivassiloglou,et al.  Predicting the Semantic Orientation of Adjectives , 1997, ACL.

[5]  Janyce Wiebe,et al.  Recognizing subjectivity: a case study in manual tagging , 1999, Natural Language Engineering.

[6]  Janyce Wiebe,et al.  Development and Use of a Gold-Standard Data Set for Subjectivity Classifications , 1999, ACL.

[7]  Janyce Wiebe,et al.  Identifying Collocations for Recognizing Opinions , 2001 .

[8]  Andrew McCallum,et al.  Conditional Random Fields: Probabilistic Models for Segmenting and Labeling Sequence Data , 2001, ICML.

[9]  Bo Pang,et al.  Thumbs up? Sentiment Classification using Machine Learning Techniques , 2002, EMNLP.

[10]  Peter D. Turney Thumbs Up or Thumbs Down? Semantic Orientation Applied to Unsupervised Classification of Reviews , 2002, ACL.

[11]  David M. Pennock,et al.  Mining the peanut gallery: opinion extraction and semantic classification of product reviews , 2003, WWW '03.

[12]  Hong Yu,et al.  Towards Answering Opinion Questions: Separating Facts from Opinions and Identifying the Polarity of Opinion Sentences , 2003, EMNLP.

[13]  Ellen Riloff,et al.  Learning Extraction Patterns for Subjective Expressions , 2003, EMNLP.

[14]  Dan Jurafsky,et al.  Automatic Extraction of Opinion Propositions and their Holders , 2004 .

[15]  Bo Pang,et al.  A Sentimental Education: Sentiment Analysis Using Subjectivity Summarization Based on Minimum Cuts , 2004, ACL.

[16]  Soo-Min Kim,et al.  Determining the Sentiment of Opinions , 2004, COLING.

[17]  Janyce Wiebe,et al.  Just How Mad Are You? Finding Strong and Weak Opinion Clauses , 2004, AAAI.

[18]  Bing Liu,et al.  Mining and summarizing customer reviews , 2004, KDD.

[19]  Yoram Singer,et al.  BoosTexter: A Boosting-based System for Text Categorization , 2000, Machine Learning.

[20]  Claire Cardie,et al.  Annotating Expressions of Opinions and Emotions in Language , 2005, Lang. Resour. Evaluation.

[21]  Shrikanth S. Narayanan,et al.  Toward detecting emotions in spoken dialogs , 2005, IEEE Transactions on Speech and Audio Processing.

[22]  Claire Cardie,et al.  Identifying Sources of Opinions with Conditional Random Fields and Extraction Patterns , 2005, HLT.

[23]  Oren Etzioni,et al.  OPINE: Extracting Product Features and Opinions from Reviews , 2005, HLT/EMNLP.

[24]  Janyce Wiebe,et al.  Recognizing Contextual Polarity in Phrase-Level Sentiment Analysis , 2005, HLT.

[25]  Eduard Hovy,et al.  Identifying Opinion Holders for Question Answering in Opinion Texts , 2005 .

[26]  Robert E. Schapire,et al.  Boosting with prior knowledge for call classification , 2005, IEEE Transactions on Speech and Audio Processing.

[27]  Oren Etzioni,et al.  Extracting Product Features and Opinions from Reviews , 2005, HLT.

[28]  Jackson Liscombe,et al.  Detecting Emotion in Speech: Experiments in Three Domains , 2006, NAACL.

[29]  Eduard Hovy,et al.  Extracting Opinions, Opinion Holders, and Topics Expressed in Online News Media Text , 2006 .

[30]  Kornel Laskowski,et al.  Emotion recognition in spontaneous speech using GMMs , 2006, INTERSPEECH.

[31]  Carolyn Penstein Rosé,et al.  Spoken Versus Typed Human and Computer Dialogue Tutoring , 2006, Int. J. Artif. Intell. Educ..

[32]  Claire Cardie,et al.  Toward Opinion Summarization: Linking the Sources , 2006 .

[33]  Frédéric Béchet,et al.  Opinion mining in a telephone survey corpus , 2006, INTERSPEECH.

[34]  Frédéric Béchet,et al.  Sequential Decision Strategies for Machine Interpretation of Speech , 2007, IEEE Transactions on Audio, Speech, and Language Processing.

[35]  Frédéric Béchet,et al.  Automatic customer feedback processing: alarm detection in open question spoken messages , 2008, INTERSPEECH.

[36]  Cynthia Whissell,et al.  The Distinct Emotional Flavor of Gnostic Writings from the Early Christian Era , 2008, Psychological reports.

[37]  Frédéric Béchet,et al.  Error correction of proportions in spoken opinion surveys , 2009, INTERSPEECH.