Efficient Crowd-Powered Active Learning for Reliable Review Evaluation

To mitigate uncertainty in the quality of online purchases (e.g., e-commerce), many people rely on review comments from others in their decision-making processes. The key challenge in this situation is how to identify useful comments among a large corpus of candidate review comments with potentially varying usefulness. In this paper, we propose the Reliable Review Evaluation Framework (RREF) which combines crowdsourcing with machine learning to address this problem. To improve crowdsourcing quality control, we propose a novel review query crowdsourcing approach which jointly considers workers' track records in review provision and current workloads when allocating review comments for workers to rate. Using the ratings crowdsourced from workers, RREF then enhances the adaptive topic classification model selection and weighting functions of AdaBoost with dynamic keyword list reconstruction. RREF has been compared with state-of-the-art related frameworks using a large-scale real-world dataset, and demonstrated over 50% reduction in average classification errors.

[1]  Karl J. Friston,et al.  Evidence for surprise minimization over value maximization in choice behavior , 2015, Scientific Reports.

[2]  Hector Garcia-Molina,et al.  Turkalytics: analytics for human computation , 2011, WWW.

[3]  Bernardete Ribeiro,et al.  On using crowdsourcing and active learning to improve classification performance , 2011, 2011 11th International Conference on Intelligent Systems Design and Applications.

[4]  Michael I. Jordan,et al.  Latent Dirichlet Allocation , 2001, J. Mach. Learn. Res..

[5]  Yoav Freund,et al.  A decision-theoretic generalization of on-line learning and an application to boosting , 1995, EuroCOLT.

[6]  Chunyan Miao,et al.  Filtering trust opinions through reinforcement learning , 2014, Decis. Support Syst..

[7]  Chunyan Miao,et al.  RepRev: Mitigating the Negative Effects of Misreported Ratings , 2014, AAAI.

[8]  Pierre Senellart,et al.  Crowd mining , 2013, SIGMOD '13.

[9]  Ido Dagan,et al.  Crowdsourcing Inference-Rule Evaluation , 2012, ACL.

[10]  Zhi-Hua Zhou,et al.  Active Learning from Crowds with Unsure Option , 2015, IJCAI.

[11]  Manuel Blum,et al.  reCAPTCHA: Human-Based Character Recognition via Web Security Measures , 2008, Science.

[12]  Steven Schockaert,et al.  Inducing semantic relations from conceptual spaces: A data-driven approach to plausible reasoning , 2015, Artif. Intell..

[13]  Pietro Michelucci,et al.  The power of crowds , 2016, Science.

[14]  Jennifer G. Dy,et al.  Active Learning from Crowds , 2011, ICML.

[15]  Raphael Silberzahn,et al.  Crowdsourced research: Many hands make tight work , 2015, Nature.

[16]  Han Su,et al.  CrowdPlanner: A Crowd-Based Route Recommendation System , 2013, ArXiv.

[17]  Chunyan Miao,et al.  Active Crowdsourcing for Annotation , 2015, 2015 IEEE/WIC/ACM International Conference on Web Intelligence and Intelligent Agent Technology (WI-IAT).

[18]  Yang Liu,et al.  A Context-Aware Topic Model for Statistical Machine Translation , 2015, ACL.

[19]  Chunyan Miao,et al.  Efficient Task Sub-Delegation for Crowdsourcing , 2015, AAAI.

[20]  Chunyan Miao,et al.  A Social Trust Model Considering Trustees' Influence , 2014, PRIMA.

[21]  Chunyan Miao,et al.  A Survey of Multi-Agent Trust Management Systems , 2013, IEEE Access.

[22]  Chunyan Miao,et al.  Challenges and Opportunities for Trust Management in Crowdsourcing , 2012, 2012 IEEE/WIC/ACM International Conferences on Web Intelligence and Intelligent Agent Technology.

[23]  Dirk Hovy,et al.  Experiments with crowdsourced re-annotation of a POS tagging data set , 2014, ACL.

[24]  Munindar P. Singh,et al.  Formal Trust Model for Multiagent Systems , 2007, IJCAI.