Exploiting Burstiness in Reviews for Review Spammer Detection

Online product reviews have become an important source of user opinions. Due to profit or fame, imposters have been writing deceptive or fake reviews to promote and/or to demote some target products or services. Such imposters are called review spammers. In the past few years, several approaches have been proposed to deal with the problem. In this work, we take a different approach, which exploits the burstiness nature of reviews to identify review spammers. Bursts of reviews can be either due to sudden popularity of products or spam attacks. Reviewers and reviews appearing in a burst are often related in the sense that spammers tend to work with other spammers and genuine reviewers tend to appear together with other genuine reviewers. This paves the way for us to build a network of reviewers appearing in different bursts. We then model reviewers and their cooccurrence in bursts as a Markov Random Field (MRF), and employ the Loopy Belief Propagation (LBP) method to infer whether a reviewer is a spammer or not in the graph. We also propose several features and employ feature induced message passing in the LBP framework for network inference. We further propose a novel evaluation method to evaluate the detected spammers automatically using supervised classification of their reviews. Additionally, we employ domain experts to perform a human evaluation of the identified spammers and non-spammers. Both the classification result and human evaluation result show that the proposed method outperforms strong baselines, which demonstrate the effectiveness of the method.

[1]  Philip S. Yu,et al.  Review spam detection via temporal pattern discovery , 2012, KDD.

[2]  Arjun Mukherjee,et al.  What Yelp Fake Review Filter Might Be Doing? , 2013, ICWSM.

[3]  J. Simonoff Multivariate Density Estimation , 1996 .

[4]  Ee-Peng Lim,et al.  Finding unusual review patterns using unexpected rules , 2010, CIKM.

[5]  Shigeo Abe DrEng Pattern Classification , 2001, Springer London.

[6]  Arjun Mukherjee,et al.  Spotting fake reviewer groups in consumer reviews , 2012, WWW.

[7]  Jiawei Han,et al.  Survey on web spam detection: principles and algorithms , 2012, SKDD.

[8]  K. Fernow New York , 1896, American Potato Journal.

[9]  Philip S. Yu,et al.  Review Graph Based Online Store Review Spammer Detection , 2011, 2011 IEEE 11th International Conference on Data Mining.

[10]  Ee-Peng Lim,et al.  Detecting product review spammers using rating behaviors , 2010, CIKM.

[11]  Christos Faloutsos,et al.  Netprobe: a fast and scalable system for fraud detection in online auction networks , 2007, WWW '07.

[12]  Claire Cardie,et al.  Finding Deceptive Opinion Spam by Any Stretch of the Imagination , 2011, ACL.

[13]  William T. Freeman,et al.  Understanding belief propagation and its generalizations , 2003 .

[14]  Yejin Choi,et al.  Distributional Footprints of Deceptive Product Reviews , 2012, ICWSM.

[15]  Bing Liu,et al.  Opinion spam and analysis , 2008, WSDM '08.

[16]  David G. Stork,et al.  Pattern Classification , 1973 .

[17]  Wolfgang Nejdl,et al.  MailRank: using ranking for spam detection , 2005, CIKM '05.

[18]  Des Spence The consumerist , 2009, BMJ : British Medical Journal.

[19]  Yi Yang,et al.  Learning to Identify Review Spam , 2011, IJCAI.

[20]  J. Fleiss Measuring nominal scale agreement among many raters. , 1971 .

[21]  Tim Oates,et al.  Detecting Spam Blogs: A Machine Learning Approach , 2006, AAAI.