From labor to trader: opinion elicitation via online crowds as a market

We often care about people's degrees of belief about certain events: e.g. causality between an action and the outcomes, odds distribution among the outcome of a horse race and so on. It is well recognized that the best form to elicit opinion from human is probability distribution instead of simple voting, because the form of distribution retains the delicate information that an opinion expresses. In the past, opinion elicitation has relied on experts, who are expensive and not always available. More recently, crowdsourcing has gained prominence as an inexpensive way to get a great deal of human input. However, traditional crowdsourcing has primarily focused on issuing very simple (e.g. binary decision) tasks to the crowd. In this paper, we study how to use crowds for Opinion Elicitation. There are three major challenges to eliciting opinion information in the form of probability distributions: how to measure the quality of distribution; how to aggregate the distributions; and, how to strategically implement such a system. To address these challenges, we design and implement COPE Crowd-powered OPinion Elicitation market. COPE models crowdsourced work as a trading market, where the "workers" behave like "traders" to maximize their profit by presenting their opinion. Among the innovative features in this system, we design COPE updating to combine the multiple elicited distributions following a Bayesian scheme. Also to provide more flexibility while running COPE, we propose a series of efficient algorithms and a slope based strategy to manage the ending condition of COPE. We then demonstrate the implementation of COPE and report experimental results running on real commercial platform to demonstrate the practical value of this system.

[1]  Christian Genest,et al.  Modeling Expert Judgments for Bayesian Updating , 1985 .

[2]  Jennifer Wortman Vaughan,et al.  Designing Informative Securities , 2012, UAI.

[3]  Gang Chen,et al.  An online cost sensitive decision-making method in crowdsourcing systems , 2013, SIGMOD '13.

[4]  Kaplan,et al.  ‘Combining Probability Distributions from Experts in Risk Analysis’ , 2000, Risk analysis : an official publication of the Society for Risk Analysis.

[5]  Tim Kraska,et al.  CrowdDB: answering queries with crowdsourcing , 2011, SIGMOD '11.

[6]  Nir Friedman,et al.  Probabilistic Graphical Models - Principles and Techniques , 2009 .

[7]  Flemming Topsøe,et al.  Some inequalities for information divergence and related measures of discrimination , 2000, IEEE Trans. Inf. Theory.

[8]  Huaiyu Zhu On Information and Sufficiency , 1997 .

[9]  R. L. Winkler,et al.  Unanimity and compromise among probability forecasters , 1990 .

[10]  Thomas M. Cover,et al.  Elements of Information Theory , 2005 .

[11]  Hisashi Kashima,et al.  Statistical quality estimation for general crowdsourcing tasks , 2013, HCOMP.

[12]  D. Helbing,et al.  How social influence can undermine the wisdom of crowd effect , 2011, Proceedings of the National Academy of Sciences.

[13]  William T. Ziemba,et al.  Efficiency of Racetrack Betting Markets , 2008 .

[14]  Jennifer Wortman Vaughan,et al.  Cost function market makers for measurable spaces , 2013, EC '13.

[15]  John Haigh,et al.  Probabilistic Risk Analysis: Foundations and Methods , 2003 .

[16]  David M. Pennock,et al.  The Real Power of Artificial Markets , 2001, Science.

[17]  Laura A. Dabbish,et al.  Labeling images with a computer game , 2004, AAAI Spring Symposium: Knowledge Collection from Volunteer Contributors.

[18]  Cyrus Shahabi,et al.  GeoCrowd: enabling query answering with spatial crowdsourcing , 2012, SIGSPATIAL/GIS.

[19]  L. J. Savage Elicitation of Personal Probabilities and Expectations , 1971 .

[20]  Robin Hanson,et al.  Combinatorial Information Market Design , 2003, Inf. Syst. Frontiers.

[21]  G. Brier VERIFICATION OF FORECASTS EXPRESSED IN TERMS OF PROBABILITY , 1950 .

[22]  Thomas A. Rietz,et al.  Accuracy and Forecast Standard Error of Prediction Markets , 2003 .

[23]  A. H. Murphy,et al.  Scoring rules and the evaluation of probabilities , 1996 .

[24]  Ted S. Sindlinger,et al.  Crowdsourcing: Why the Power of the Crowd is Driving the Future of Business , 2010 .

[25]  Jacob D. Abernethy,et al.  A Collaborative Mechanism for Crowdsourcing Prediction Problems , 2011, NIPS.

[26]  E. H. Shuford,et al.  Admissible probability measurement procedures , 1966, Psychometrika.

[27]  Christian Genest,et al.  Combining Probability Distributions: A Critique and an Annotated Bibliography , 1986 .

[28]  Ian A. Kash,et al.  Information elicitation for decision making , 2011, AAMAS.

[29]  Thomas Pfeiffer,et al.  Adaptive Polling for Information Aggregation , 2012, AAAI.

[30]  David R. Karger,et al.  Counting with the Crowd , 2012, Proc. VLDB Endow..

[31]  Lei Chen,et al.  Whom to Ask? Jury Selection for Decision Making Tasks on Micro-blog Services , 2012, Proc. VLDB Endow..

[32]  Pierre Senellart,et al.  Crowd mining , 2013, SIGMOD '13.