Task assignment optimization in knowledge-intensive crowdsourcing

We present SmartCrowd, a framework for optimizing task assignment in knowledge-intensive crowdsourcing (KI-C). SmartCrowd distinguishes itself by formulating, for the first time, the problem of worker-to-task assignment in KI-C as an optimization problem, by proposing efficient adaptive algorithms to solve it and by accounting for human factors, such as worker expertise, wage requirements, and availability inside the optimization process. We present rigorous theoretical analyses of the task assignment optimization problem and propose optimal and approximation algorithms with guarantees, which rely on index pre-computation and adaptive maintenance. We perform extensive performance and quality experiments using real and synthetic data to demonstrate that the SmartCrowd approach is necessary to achieve efficient task assignments of high-quality under guaranteed cost budget.

[1]  Jennifer Widom,et al.  CrowdScreen: algorithms for filtering data with humans , 2012, SIGMOD Conference.

[2]  David S. Johnson,et al.  Computers and Intractability: A Guide to the Theory of NP-Completeness , 1978 .

[3]  Vahab Mirrokni,et al.  Maximizing Non-Monotone Submodular Functions , 2007, FOCS 2007.

[4]  Haim Kaplan,et al.  Answering Planning Queries with the Crowd (Technical report) , 2013 .

[5]  Tim Kraska,et al.  CrowdER: Crowdsourcing Entity Resolution , 2012, Proc. VLDB Endow..

[6]  Seog Park,et al.  A quality enhancement of crowdsourcing based on quality evaluation and user-level task assignment framework , 2014, 2014 International Conference on Big Data and Smart Computing (BIGCOMP).

[7]  Luca Becchetti,et al.  Online team formation in social networks , 2012, WWW.

[8]  Despina Polemi,et al.  CorpWiki: A self-regulating wiki to promote corporate collective intelligence through expert peer matching , 2010, Inf. Sci..

[9]  Jennifer Widom,et al.  Query Optimization over Crowdsourced Data , 2013, Proc. VLDB Endow..

[10]  B. H. Rudall Lecture notes in computer science. Vol. 15—L-systems , 1978 .

[11]  M. Hossain,et al.  Crowdsourcing: Activities, incentives and users' motivations to participate , 2012, 2012 International Conference on Innovation Management and Technology Research.

[12]  Klaudia Frankfurter Computers And Intractability A Guide To The Theory Of Np Completeness , 2016 .

[13]  Kwong-Sak Leung,et al.  Task recommendation in crowdsourcing systems , 2012, CrowdKDD '12.

[14]  Ohad Greenshpan,et al.  Asking the Right Questions in Crowd Data Sourcing , 2012, 2012 IEEE 28th International Conference on Data Engineering.

[15]  Tim Kraska,et al.  CrowdDB: Query Processing with the VLDB Crowd , 2011, Proc. VLDB Endow..

[16]  Björn Hartmann,et al.  Collaboratively crowdsourcing workflows with turkomatic , 2012, CSCW.

[17]  Hisashi Kashima,et al.  Statistical quality estimation for general crowdsourcing tasks , 2013, HCOMP.

[18]  Daniel J. Veit,et al.  More than fun and money. Worker Motivation in Crowdsourcing - A Study on Mechanical Turk , 2011, AMCIS.

[19]  David R. Karger,et al.  Counting with the Crowd , 2012, Proc. VLDB Endow..

[20]  R. Suganya,et al.  Data Mining Concepts and Techniques , 2010 .

[21]  Devavrat Shah,et al.  Budget-Optimal Task Allocation for Reliable Crowdsourcing Systems , 2011, Oper. Res..

[22]  Edilaine Martins Soler,et al.  A modified Primal-Dual Logarithmic-Barrier Method for solving the Optimal Power Flow problem with discrete and continuous control variables , 2012, Eur. J. Oper. Res..

[23]  Pável Calado,et al.  Automatic Assessment of Document Quality in Web Collaborative Digital Libraries , 2011, JDIQ.

[24]  Hisashi Kashima,et al.  Crowdsourcing Quality Control for Item Ordering Tasks , 2013, HCOMP.

[25]  Vahab S. Mirrokni,et al.  Maximizing Non-Monotone Submodular Functions , 2011, 48th Annual IEEE Symposium on Foundations of Computer Science (FOCS'07).

[26]  Vahab S. Mirrokni,et al.  Maximizing Nonmonotone Submodular Functions under Matroid or Knapsack Constraints , 2009, SIAM J. Discret. Math..

[27]  Aniket Kittur,et al.  Coordination in collective intelligence: the role of team structure and task interdependence , 2009, CHI.

[28]  Tim Kraska,et al.  Leveraging transitive relations for crowdsourced joins , 2013, SIGMOD '13.

[29]  Aditya Ramesh Identifying Reliable Workers Swiftly , 2012 .

[30]  Jan Vondr Symmetry and Approximability of Submodular Maximization Problems , 2013 .

[31]  Eric Horvitz,et al.  Crowdsourcing General Computation , 2011 .

[32]  Javier R. Movellan,et al.  Whose Vote Should Count More: Optimal Integration of Labels from Labelers of Unknown Expertise , 2009, NIPS.

[33]  Gagan Goel,et al.  Allocating tasks to workers with matching constraints: truthful mechanisms for crowdsourcing markets , 2014, WWW.

[34]  M. L. Fisher,et al.  An analysis of approximations for maximizing submodular set functions—I , 1978, Math. Program..

[35]  Paola Alimonti Non-oblivious Local Search for MAX 2-CCSP with Application to MAX DICUT , 1997, WG.

[36]  Lorrie Faith Cranor,et al.  Are your participants gaming the system?: screening mechanical turk workers , 2010, CHI.

[37]  David R. Karger,et al.  Human-powered Sorts and Joins , 2011, Proc. VLDB Endow..

[38]  Aditya G. Parameswaran,et al.  So who won?: dynamic max discovery with the crowd , 2012, SIGMOD Conference.

[39]  Brian S. Butler,et al.  Rules and Roles vs. Consensus , 2013 .

[40]  Tharam S. Dillon,et al.  Content Quality Assessment Related Frameworks for Social Media , 2009, ICCSA.

[41]  Panagiotis G. Ipeirotis,et al.  Quizz: targeted crowdsourcing with a billion (potential) users , 2014, WWW.

[42]  Panagiotis G. Ipeirotis,et al.  Quality management on Amazon Mechanical Turk , 2010, HCOMP '10.

[43]  Beth A. Bechky,et al.  The Emergence of Governance in an Open Source Community , 2007 .

[44]  Yannick Naudet,et al.  Improving Wiki Article Quality Through Crowd Coordination: A Resource Allocation Approach , 2013, Int. J. Semantic Web Inf. Syst..

[45]  Haim Kaplan,et al.  Answering Planning Queries with the Crowd , 2013, Proc. VLDB Endow..

[46]  Scott R. Klemmer,et al.  Shepherding the crowd yields better work , 2012, CSCW.

[47]  John Riedl,et al.  Is Wikipedia growing a longer tail? , 2009, GROUP.

[48]  Sihem Amer-Yahia,et al.  Crowds, not Drones: Modeling Human Factors in Interactive Crowdsourcing , 2013, DBCrowd.

[49]  Peng Dai,et al.  Inserting Micro-Breaks into Crowdsourcing Workflows , 2013, HCOMP.

[50]  Audun Jøsang,et al.  A survey of trust and reputation systems for online service provision , 2007, Decis. Support Syst..

[51]  Mausam,et al.  Crowdsourcing Multi-Label Classification for Taxonomy Creation , 2013, HCOMP.

[52]  Aniket Kittur,et al.  A comparison of social, learning, and financial strategies on crowd engagement and output quality , 2014, CSCW.

[53]  Michael S. Bernstein,et al.  The future of crowd work , 2013, CSCW.

[54]  Chien-Ju Ho,et al.  Online Task Assignment in Crowdsourcing Markets , 2012, AAAI.

[55]  Dana Chandler,et al.  Breaking Monotony with Meaning: Motivation in Crowdsourcing Markets , 2012, ArXiv.