Elicitation and Aggregation of Crowd Information

This thesis addresses challenges in elicitation and aggregation of crowd information for settings where an information collector, called center, has a limited knowledge about information providers, called agents. Each agent is assumed to have noisy private information that brings a high information gain to the center when it is aggregated with the private information of other agents. We address two particular issues in eliciting crowd information: 1) how to incentivize agents to participate and provide accurate data; 2) how to aggregate crowd information so that the negative impact of agents who provide low quality information is bounded. We examine three different information elicitation settings. In the first elicitation setting, agents report their observations regarding a single phenomenon that represents an abstraction of a crowdsourcing task. The center itself does not observe the phenomenon, so it rewards agents by comparing their reports. Clearly, a rational agent bases her reporting strategy on what she believes about other agents, called peers. We prove that, in general, no payment mechanism can achieve strict properness (i.e., adopt truthful reporting as a strict equilibrium strategy) if agents only report their observations, even if they share a common belief system. This motivates the use of payment mechanisms that are based on an additional report. We show that a general payment mechanism cannot have a simple structure, often adopted by prior work, and that in the limit case, when observations can take real values, agents are constrained to share a common belief system. Furthermore, we develop several payment mechanisms for the elicitation of non-binary observations. In the second elicitation setting, a group of agents observes multiple a priori similar phenomena. Due to the a priori similarity condition, the setting represents a refinement of the former setting and enables one to achieve stronger incentive properties without requiring additional reports or constraining agents to share a common belief system. We extend the existing mechanisms to allow non-binary observations by constructing strongly truthful mechanisms (i.e., mechanisms in which truthful reporting is the highest-paying equilibrium) for different types of agents' population. In the third elicitation setting, agents observe a time evolving phenomenon, and a few of them, whose identity is known, are trusted to report truthful observations. The existence of trusted agents makes this setting much more stringent than the previous ones. We show that, in the context of online information aggregation, one can not only incentivize agents to provide informative reports, but also limit the effectiveness of malicious agents who deliberately misreport. To do so, we construct a reputation system that puts a bound on the negative impact that any misreporting strategy can have on the learned aggregate. Finally, we experimentally verify the effectiveness of novel elicitation mechanisms in community sensing simulation testbeds and a peer grading experiment.

[1]  Robin Hanson,et al.  Combinatorial Information Market Design , 2003, Inf. Syst. Frontiers.

[2]  Nicholas R. Jennings,et al.  Trust-based fusion of untrustworthy information in crowdsourcing applications , 2013, AAMAS.

[3]  Arpit Agarwal,et al.  Informed Truthfulness in Multi-Task Peer Prediction , 2016, EC.

[4]  G. Brier VERIFICATION OF FORECASTS EXPRESSED IN TERMS OF PROBABILITY , 1950 .

[5]  Wen Hu,et al.  On the need for a reputation system in mobile phone based sensing , 2014, Ad Hoc Networks.

[6]  Lakshminarayanan Subramanian,et al.  Reputation-based Worker Filtering in Crowdsourcing , 2014, NIPS.

[7]  Gábor Lugosi,et al.  Prediction, learning, and games , 2006 .

[8]  Wen Hu,et al.  Towards trustworthy participatory sensing , 2009 .

[9]  A. Raftery,et al.  Strictly Proper Scoring Rules, Prediction, and Estimation , 2007 .

[10]  G. Loewenstein,et al.  Measuring the Prevalence of Questionable Research Practices With Incentives for Truth Telling , 2012, Psychological science.

[11]  Ryan P. Adams,et al.  Trick or treat: putting peer prediction to the test , 2014 .

[12]  Zhenghao Chen,et al.  Tuned Models of Peer Assessment in MOOCs , 2013, EDM.

[13]  David M. Pennock,et al.  Collective revelation: a mechanism for self-verified, weighted, and truthful predictions , 2009, EC '09.

[14]  D. Prelec A Bayesian Truth Serum for Subjective Data , 2004, Science.

[15]  Drazen Prelec,et al.  Creating Truth-Telling Incentives with the Bayesian Truth Serum , 2013 .

[16]  Yoav Shoham,et al.  Multiagent Systems - Algorithmic, Game-Theoretic, and Logical Foundations , 2009 .

[17]  Christopher K. I. Williams,et al.  Gaussian Processes for Machine Learning (Adaptive Computation and Machine Learning) , 2005 .

[18]  Boi Faltings,et al.  A Robust Bayesian Truth Serum for Non-Binary Signals , 2013, AAAI.

[19]  Boi Faltings,et al.  Minimum payments that reward honest reputation feedback , 2006, EC '06.

[20]  Paul Resnick,et al.  Eliciting Informative Feedback: The Peer-Prediction Method , 2005, Manag. Sci..

[21]  Richard P. McLean,et al.  Optimal Selling Strategies under Uncertainty for a Discriminating Monopolist When Demands Are Interdependent , 1985 .

[22]  Mihaela van der Schaar,et al.  Socially-Optimal Design of Service Exchange Platforms with Imperfect Monitoring , 2015, ACM Trans. Economics and Comput..

[23]  Boi Faltings,et al.  Incentives for Subjective Evaluations with Private Beliefs , 2015, AAAI.

[24]  Manfred K. Warmuth,et al.  The Weighted Majority Algorithm , 1994, Inf. Comput..

[25]  Pramod K. Varshney,et al.  Reliable Crowdsourcing for Multi-Class Labeling Using Coding Theory , 2013, IEEE Journal of Selected Topics in Signal Processing.

[26]  Kevin Leyton-Brown,et al.  Mechanical TA: Partially Automated High-Stakes Peer Grading , 2015, SIGCSE.

[27]  Yoav Shoham,et al.  Eliciting properties of probability distributions , 2008, EC '08.

[28]  Avinash Srinivasan,et al.  iTrust: an integrated trust framework for wireless sensor networks , 2010, SAC '10.

[29]  David C. Parkes,et al.  Peer prediction without a common prior , 2012, EC '12.

[30]  Jens Witkowski Robust peer prediction mechanisms , 2015 .

[31]  Eric Horvitz,et al.  Incentives for truthful reporting in crowdsourcing , 2012, AAMAS.

[32]  Boi Faltings,et al.  Incentives for Truthful Information Elicitation of Continuous Signals , 2014, AAAI.

[33]  K. Scherer What are emotions? And how can they be measured? , 2005 .

[34]  G. Radanovic,et al.  Neural network based speech synthesis for Croatian language , 2010, The 33rd International Convention MIPRO.

[35]  Boi Faltings,et al.  Incentives for expressing opinions in online polls , 2008, EC '08.

[36]  Boi Faltings,et al.  Incentive Schemes for Participatory Sensing , 2015, AAMAS.

[37]  Landon P. Cox,et al.  YouProve: authenticity and fidelity in mobile sensing , 2011, SenSys.

[38]  Justin Cheng,et al.  Peer and self assessment in massive online classes , 2013, ACM Trans. Comput. Hum. Interact..

[39]  S. Buchegger,et al.  A Robust Reputation System for Mobile Ad-hoc Networks , 2003 .

[40]  Laura A. Dabbish,et al.  Labeling images with a computer game , 2004, AAAI Spring Symposium: Knowledge Collection from Volunteer Contributors.

[41]  Paul Resnick,et al.  The influence limiter: provably manipulation-resistant recommender systems , 2007, RecSys '07.

[42]  Boi Faltings,et al.  Mechanisms for Making Crowds Truthful , 2014, J. Artif. Intell. Res..

[43]  Wai-Tat Fu,et al.  Don't hide in the crowd!: increasing social transparency between peer workers improves crowdsourcing outcomes , 2013, CHI.

[44]  Kevin Leyton-Brown,et al.  Incentivizing Evaluation via Limited Access to Ground Truth: Peer-Prediction Makes Things Worse , 2016, ArXiv.

[45]  Roy N. Colvile,et al.  Uncertainty in dispersion modelling and urban air quality mapping , 2002 .

[46]  L. A. Goodman,et al.  Social Choice and Individual Values , 1951 .

[47]  David C. Parkes,et al.  A Robust Bayesian Truth Serum for Small Populations , 2012, AAAI.

[48]  Aaron D. Shaw,et al.  Designing incentives for inexpert human raters , 2011, CSCW.

[49]  David C. Parkes,et al.  Measuring Performance of Peer Prediction Mechanisms Using Replicator Dynamics , 2016, IJCAI.

[50]  Grant Schoenebeck,et al.  Equilibrium Selection in Information Elicitation without Verification via Information Monotonicity , 2018, ITCS.

[51]  David M. Pennock,et al.  The Real Power of Artificial Markets , 2001, Science.

[52]  Boi Faltings,et al.  A Region-Based Model for Estimating Urban Air Pollution , 2014, AAAI.

[53]  Devavrat Shah,et al.  Efficient crowdsourcing for multi-class labeling , 2013, SIGMETRICS '13.

[54]  Mihaela van der Schaar,et al.  Reputation-based incentive protocols in crowdsourcing applications , 2011, 2012 Proceedings IEEE INFOCOM.

[55]  L. J. Savage Elicitation of Personal Probabilities and Expectations , 1971 .

[56]  Jens Witkowski,et al.  A Geometric Method to Construct Minimal Peer Prediction Mechanisms , 2016, AAAI.

[57]  Boi Faltings,et al.  Incentives to Counter Bias in Human Computation , 2014, HCOMP.

[58]  Yiling Chen,et al.  Elicitability and knowledge-free elicitation with peer prediction , 2014, AAMAS.

[59]  Boi Faltings,et al.  Incentivizing truthful responses with the logarithmic peer truth serum , 2015, UbiComp/ISWC Adjunct.

[60]  Steven Reece,et al.  Multi-sensor fault recovery in the presence of known and unknown fault types , 2009, 2009 12th International Conference on Information Fusion.

[61]  Gerardo Hermosillo,et al.  Learning From Crowds , 2010, J. Mach. Learn. Res..

[62]  Lukas Biewald,et al.  Programmatic Gold: Targeted and Scalable Quality Assurance in Crowdsourcing , 2011, Human Computation.

[63]  Tim Roughgarden,et al.  Algorithmic Game Theory , 2007 .

[64]  Michael Vitale,et al.  The Wisdom of Crowds , 2015, Cell.

[65]  Yoav Shoham,et al.  Truthful Surveys , 2008, WINE.

[66]  Refik Molva,et al.  Core: a collaborative reputation mechanism to enforce node cooperation in mobile ad hoc networks , 2002, Communications and Multimedia Security.

[67]  David M. Pennock,et al.  A Utility Framework for Bounded-Loss Market Makers , 2007, UAI.

[68]  Boi Faltings,et al.  Incentives for Effort in Crowdsourcing Using the Peer Truth Serum , 2016, ACM Trans. Intell. Syst. Technol..

[69]  Andreas Krause,et al.  Incentives for Privacy Tradeoff in Community Sensing , 2013, HCOMP.

[70]  Lothar Thiele,et al.  OpenSense: open community driven sensing of environment , 2010, IWGS '10.

[71]  M. Kearns,et al.  An Algorithm That Finds Truth Even If Most People Are Wrong , 2007 .

[72]  Boi Faltings,et al.  Limiting the Influence of Low Quality Information in Community Sensing , 2016, AAMAS.

[73]  Shunsuke Ihara,et al.  Information theory - for continuous systems , 1993 .

[74]  Jennifer Wortman Vaughan,et al.  A new understanding of prediction markets via no-regret learning , 2010, EC '10.

[75]  Salil S. Kanhere,et al.  A survey on privacy in mobile participatory sensing applications , 2011, J. Syst. Softw..

[76]  D. Helbing,et al.  How social influence can undermine the wisdom of crowd effect , 2011, Proceedings of the National Academy of Sciences.

[77]  Boi Faltings,et al.  Swissnoise: Online Polls with Game-Theoretic Incentives , 2014, AAAI.

[78]  Boi Faltings,et al.  Incentives for Answering Hypothetical Questions , 2011 .

[79]  Arpita Ghosh,et al.  Game Theory and Incentives in Human Computation Systems , 2013, Handbook of Human Computation.

[80]  Audun Jøsang,et al.  AIS Electronic Library (AISeL) , 2017 .

[81]  Anirban Dasgupta,et al.  Crowdsourced judgement elicitation with endogenous proficiency , 2013, WWW.

[82]  Yiling Chen,et al.  Output Agreement Mechanisms and Common Knowledge , 2014, HCOMP.

[83]  Grant Schoenebeck,et al.  G T ] 3 M ay 2 01 6 A Framework For Designing Information Elicitation Mechanisms That Reward Truth-telling Yuqing Kong University of Michigan , 2016 .

[84]  Nihar B. Shah,et al.  Double or Nothing: Multiplicative Incentive Mechanisms for Crowdsourcing , 2014, J. Mach. Learn. Res..

[85]  Boi Faltings,et al.  Incentive Mechanisms for Community Sensing , 2014, IEEE Transactions on Computers.

[86]  Bin Bi,et al.  Iterative Learning for Reliable Crowdsourcing Systems , 2012 .

[87]  Boi Faltings,et al.  Eliciting Truthful Information with the Peer Truth Serum , 2014 .

[88]  Yoav Shoham,et al.  Eliciting truthful answers to multiple-choice questions , 2009, EC '09.

[89]  Boi Faltings,et al.  Robust Incentive-Compatible Feedback Payments , 2006, TADA/AMEC.

[90]  Luca de Alfaro,et al.  Incentives for Truthful Peer Grading , 2015, ArXiv.

[91]  Peter Auer,et al.  Finite-time Analysis of the Multiarmed Bandit Problem , 2002, Machine Learning.

[92]  R. Sami,et al.  Bounded-Regret Sequential Learning using Prediction Markets ⇤ ( Extended Abstract ) , 2011 .

[93]  Arpita Ghosh,et al.  Optimal Contest Design for Simple Agents , 2016, TEAC.

[94]  Mani B. Srivastava,et al.  Reputation-based framework for high integrity sensor networks , 2004, SASN '04.

[95]  Chrysanthos Dellarocas,et al.  Reputation Mechanism Design in Online Trading Environments with Pure Moral Hazard , 2005, Inf. Syst. Res..

[96]  Kannan Ramchandran,et al.  Truth Serums for Massively Crowdsourced Evaluation Tasks , 2015, ArXiv.

[97]  David C. Parkes,et al.  Practical Peer Prediction for Peer Assessment , 2016, HCOMP.

[98]  Yuval Peres,et al.  Approval Voting and Incentives in Crowdsourcing , 2015, ICML.

[99]  M. Hansen,et al.  Participatory Sensing , 2019, Internet of Things.

[100]  Laura A. Dabbish,et al.  Designing games with a purpose , 2008, CACM.

[101]  Ian A. Kash,et al.  Elicitation for Aggregation , 2014, AAAI.

[102]  Andreas Krause,et al.  Near-Optimal Sensor Placements in Gaussian Processes: Theory, Efficient Algorithms and Empirical Studies , 2008, J. Mach. Learn. Res..

[103]  Haiguang Chen,et al.  Task-based Trust Management for Wireless Sensor Networks , 2009 .

[104]  Boi Faltings,et al.  Learning to Scale Payments in Crowdsourcing with PropeRBoost , 2016, HCOMP.

[105]  Christopher G. Harris You're Hired! An Examination of Crowdsourcing Incentive Models in Human Resource Tasks , 2011 .

[106]  Jian Peng,et al.  Variational Inference for Crowdsourcing , 2012, NIPS.

[107]  M. F.,et al.  Bibliography , 1985, Experimental Gerontology.

[108]  Jean-Yves Le Boudec,et al.  Performance analysis of the CONFIDANT protocol , 2002, MobiHoc '02.

[109]  Carlos Guestrin,et al.  The Wisdom of Multiple Guesses , 2015, EC.

[110]  Radford M. Neal Pattern Recognition and Machine Learning , 2007, Technometrics.

[111]  Yishay Mansour,et al.  Improved second-order bounds for prediction with expert advice , 2006, Machine Learning.

[112]  Sindhu Kutty A Prediction Market Approach to Learning with Sequential Advice , 2010 .

[113]  Lothar Thiele,et al.  Pushing the spatio-temporal resolution limit of urban air pollution maps , 2014, 2014 IEEE International Conference on Pervasive Computing and Communications (PerCom).

[114]  Aurélien Garivier,et al.  The KL-UCB Algorithm for Bounded Stochastic Bandits and Beyond , 2011, COLT.

[115]  R. Hanson LOGARITHMIC MARKETS CORING RULES FOR MODULAR COMBINATORIAL INFORMATION AGGREGATION , 2012 .

[116]  Alec Wolman,et al.  I am a sensor, and I approve this message , 2010, HotMobile '10.

[117]  Andreas Krause,et al.  Toward Community Sensing , 2008, 2008 International Conference on Information Processing in Sensor Networks (ipsn 2008).

[118]  S. Matthew Weinberg,et al.  Game Theory based Peer Grading Mechanisms for MOOCs , 2015, L@S.

[119]  Blake Riley,et al.  Minimum Truth Serums with Optional Predictions , 2014 .

[120]  Zhu Han,et al.  CatchIt: Detect Malicious Nodes in Collaborative Spectrum Sensing , 2009, GLOBECOM 2009 - 2009 IEEE Global Telecommunications Conference.

[121]  Paul A. Pavlou,et al.  Can online reviews reveal a product's true quality?: empirical findings and analytical modeling of Online word-of-mouth communication , 2006, EC '06.

[122]  David C. Parkes,et al.  Learning the Prior in Minimal Peer Prediction , 2013 .

[123]  T. L. Lai Andherbertrobbins Asymptotically Efficient Adaptive Allocation Rules , 2022 .

[124]  Claudiu Cristian Musat,et al.  EmotionWatch: Visualizing Fine-Grained Emotions in Event-Related Tweets , 2014, ICWSM.

[125]  Arpita Ghosh,et al.  Behavioral Mechanism Design: Optimal Crowdsourcing Contracts and Prospect Theory , 2015, EC.

[126]  Wai-Tat Fu,et al.  Enhancing reliability using peer consistency evaluation in human computation , 2013, CSCW '13.

[127]  Grant Schoenebeck,et al.  Putting Peer Prediction Under the Micro(economic)scope and Making Truth-Telling Focal , 2016, WINE.