Detection of opinion spam based on anomalous rating deviation

We propose a novel approach for detecting opinion spam based on reviewer ratings, with no need for text analysis.Our approach identifies reviewers having a significant number of reviews that disagree with majority opinion.Our approach requires significantly less running time than other, comparable approaches.Our approach is shown to successfully identify opinion spammers in synthetic and real world data sets. The publication of fake reviews by parties with vested interests has become a severe problem for consumers who use online product reviews in their decision making. To counter this problem a number of methods for detecting these fake reviews, termed opinion spam, have been proposed. However, to date, many of these methods focus on analysis of review text, making them unsuitable for many review systems where accompanying text is optional, or not possible. Moreover, these approaches are often computationally expensive, requiring extensive resources to handle text analysis over the scale of data typically involved.In this paper, we consider opinion spammers manipulation of average ratings for products, focusing on differences between spammer ratings and the majority opinion of honest reviewers. We propose a lightweight, effective method for detecting opinion spammers based on these differences. This method uses binomial regression to identify reviewers having an anomalous proportion of ratings that deviate from the majority opinion. Experiments on real-world and synthetic data show that our approach is able to successfully identify opinion spammers. Comparison with the current state-of-the-art approach, also based only on ratings, shows that our method is able to achieve similar detection accuracy while removing the need for assumptions regarding probabilities of spam and non-spam reviews and reducing the heavy computation required for learning.

[1]  Philip S. Yu,et al.  Review spam detection via temporal pattern discovery , 2012, KDD.

[2]  Ee-Peng Lim,et al.  Detecting product review spammers using rating behaviors , 2010, CIKM.

[3]  Claire Cardie,et al.  Finding Deceptive Opinion Spam by Any Stretch of the Imagination , 2011, ACL.

[4]  Sameer Singh,et al.  Novelty detection: a review - part 1: statistical approaches , 2003, Signal Process..

[5]  Shady Aly,et al.  Fuzzy aggregation and averaging for group decision making: A generalization and survey , 2009, Knowl. Based Syst..

[6]  Arjun Mukherjee,et al.  Spotting fake reviewer groups in consumer reviews , 2012, WWW.

[7]  Philip S. Yu,et al.  Identify Online Store Review Spammers via Social Review Graph , 2012, TIST.

[8]  Christopher Krügel,et al.  Anomaly detection of web-based attacks , 2003, CCS '03.

[9]  Christos Faloutsos,et al.  Opinion Fraud Detection in Online Reviews by Network Effects , 2013, ICWSM.

[10]  Jennifer Jie Xu,et al.  Knowledge Discovery and Data Mining , 2014, Computing Handbook, 3rd ed..

[11]  Derek Greene,et al.  Distortion as a validation criterion in the identification of suspicious reviews , 2010, SOMA '10.

[12]  RossoPaolo,et al.  Detecting positive and negative deceptive opinions using PU-learning , 2015 .

[13]  VARUN CHANDOLA,et al.  Anomaly detection: A survey , 2009, CSUR.

[14]  Naomie Salim,et al.  Detection of review spam: A survey , 2015, Expert Syst. Appl..

[15]  Logan Kugler Keeping online reviews honest , 2014, Commun. ACM.

[16]  Yi Yang,et al.  Learning to Identify Review Spam , 2011, IJCAI.

[17]  Ronald R. Yager,et al.  Modeling the concept of majority opinion in group decision making , 2006, Inf. Sci..

[18]  L. Kratz,et al.  Anomaly detection in extremely crowded scenes using spatio-temporal motion pattern models , 2009, 2009 IEEE Conference on Computer Vision and Pattern Recognition.

[19]  Paolo Rosso,et al.  Detecting positive and negative deceptive opinions using PU-learning , 2015, Inf. Process. Manag..

[20]  Raymond Y. K. Lau,et al.  Text mining and probabilistic language modeling for online review spam detection , 2012, TMIS.

[21]  Bing Liu,et al.  Opinion spam and analysis , 2008, WSDM '08.

[22]  Christos Faloutsos,et al.  RTG: a recursive realistic graph generator using random typing , 2009, Data Mining and Knowledge Discovery.

[23]  Abhinav Kumar,et al.  Spotting opinion spammers using behavioral footprints , 2013, KDD.

[24]  Arjun Mukherjee,et al.  Exploiting Burstiness in Reviews for Review Spammer Detection , 2021, ICWSM.

[25]  Damien Ernst,et al.  SPRT for SPIT: Using the Sequential Probability Ratio Test for Spam in VoIP Prevention , 2012, AIMS.

[26]  Louis Kratz,et al.  Anomaly detection in extremely crowded scenes using spatio-temporal motion pattern models , 2009, CVPR.

[27]  S. Swamynathan,et al.  Scoring products from reviews through application of fuzzy techniques , 2010, Expert Syst. Appl..

[28]  Eleazar Eskin,et al.  Anomaly Detection over Noisy Data using Learned Probability Distributions , 2000, ICML.

[29]  Raymond Y. K. Lau,et al.  Text mining and probabilistic language modeling for online review spam detecting , 2011 .