Leveraging the crowd to improve feature-sentiment analysis of user reviews

Crowdsourcing and machine learning are both useful techniques for solving difficult problems (e.g., computer vision and natural language processing). In this paper, we propose a novel method that harnesses and combines the strength of these two techniques to better analyze the features and the sentiments toward them in user reviews. To strike a good balance between reducing information overload and providing the original context expressed by review writers, the proposed system (1) allows users to interactively rank the entities based on feature-rating, (2) automatically highlights sentences that are related to relevant features, and (3) utilizes implicit crowdsourcing by encouraging users to provide correct labels of their own reviews to improve the feature-sentiment classifier. The proposed system not only helps users to save time and effort to digest the often massive amount of user reviews, but also provides real-time suggestions on relevant features and ratings as users generate their own reviews. Results from a simulation experiment show that leveraging on the crowd can significantly improve the feature-sentiment analysis of user reviews. Furthermore, results from a user study show that the proposed interface was preferred by more participants than interfaces that use traditional noun-adjective pair summarization, as the current interface allows users to view feature-related information in the original context.

[1]  Laura A. Dabbish,et al.  Designing games with a purpose , 2008, CACM.

[2]  David M. Blei,et al.  Supervised Topic Models , 2007, NIPS.

[3]  Oren Etzioni,et al.  RevMiner: an extractive interface for navigating reviews on a smartphone , 2012, UIST.

[4]  Andrea Esuli,et al.  SENTIWORDNET: A Publicly Available Lexical Resource for Opinion Mining , 2006, LREC.

[5]  W. Bruggeman,et al.  APPENDIX : QUESTIONNAIRE I , 2009 .

[6]  Bo Pang,et al.  Thumbs up? Sentiment Classification using Machine Learning Techniques , 2002, EMNLP.

[7]  Huaiyu Zhu On Information and Sufficiency , 1997 .

[8]  Daphne Koller,et al.  Support Vector Machine Active Learning with Applications to Text Classification , 2000, J. Mach. Learn. Res..

[9]  Yue Lu,et al.  Latent aspect rating analysis on review text data: a rating regression approach , 2010, KDD.

[10]  Alexander Dekhtyar,et al.  Information Retrieval , 2018, Lecture Notes in Computer Science.

[11]  Hao Su,et al.  Crowdsourcing Annotations for Visual Object Detection , 2012, HCOMP@AAAI.

[12]  Jane Yung-jen Hsu,et al.  KissKissBan: a competitive human computation game for image annotation , 2010, HCOMP '09.

[13]  Peter D. Turney Thumbs Up or Thumbs Down? Semantic Orientation Applied to Unsupervised Classification of Reviews , 2002, ACL.

[14]  Umesh V. Vazirani,et al.  An Introduction to Computational Learning Theory , 1994 .

[15]  Cyrus Rashtchian,et al.  Collecting Image Annotations Using Amazon’s Mechanical Turk , 2010, Mturk@HLT-NAACL.

[16]  Giuseppe Carenini,et al.  A multimedia interface for facilitating comparisons of opinions , 2009, IUI.

[17]  Brendan T. O'Connor,et al.  Cheap and Fast – But is it Good? Evaluating Non-Expert Annotations for Natural Language Tasks , 2008, EMNLP.

[18]  Giuseppe Carenini,et al.  Interactive multimedia summaries of evaluative text , 2006, IUI '06.

[19]  David A. Forsyth,et al.  Utility data annotation with Amazon Mechanical Turk , 2008, 2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops.

[20]  Manuel Blum,et al.  Peekaboom: a game for locating objects in images , 2006, CHI.

[21]  Koji Yatani,et al.  Review spotlight: a user interface for summarizing user-generated reviews using adjective-noun word pairs , 2011, CHI.

[22]  David M. Pennock,et al.  Mining the peanut gallery: opinion extraction and semantic classification of product reviews , 2003, WWW '03.

[23]  Bing Liu,et al.  Opinion observer: analyzing and comparing opinions on the Web , 2005, WWW '05.

[24]  Michael S. Bernstein,et al.  Soylent: a word processor with a crowd inside , 2010, UIST.

[25]  Laura A. Dabbish,et al.  Labeling images with a computer game , 2004, AAAI Spring Symposium: Knowledge Collection from Volunteer Contributors.

[26]  Barry Smyth,et al.  Towards an intelligent reviewer's assistant: recommending topics to help users to write better product reviews , 2012, IUI '12.

[27]  Rob Miller,et al.  VizWiz: nearly real-time answers to visual questions , 2010, UIST.

[28]  Jeffrey Nichols,et al.  Summarizing sporting events using twitter , 2012, IUI '12.

[29]  Chris Callison-Burch,et al.  Creating Speech and Language Data With Amazon’s Mechanical Turk , 2010, Mturk@HLT-NAACL.

[30]  Bing Liu,et al.  Mining and summarizing customer reviews , 2004, KDD.