Partial Information Framework: Aggregating Estimates from Diverse Information Sources

Prediction polling is an increasingly popular form of crowdsourcing in which multiple participants estimate the probability or magnitude of some future event. These estimates are then aggregated into a single forecast. Historically, randomness in scientific estimation has been generally assumed to arise from unmeasured factors which are viewed as measurement noise. However, when combining subjective estimates, heterogeneity stemming from differences in the participants' information is often more important than measurement noise. This paper formalizes information diversity as an alternative source of such heterogeneity and introduces a novel modeling framework that is particularly well-suited for prediction polls. A practical specification of this framework is proposed and applied to the task of aggregating probability and point estimates from two real-world prediction polls. In both cases our model outperforms standard measurement-error-based aggregators, hence providing evidence in favor of information diversity being the more important source of heterogeneity.

[1]  F. Mosteller,et al.  Low Moments for Small Samples: A Comparative Study of Order Statistics , 1947 .

[2]  Boris Polyak,et al.  The method of projections for finding the common point of convex sets , 1967 .

[3]  A. Dawid The Well-Calibrated Bayesian , 1982 .

[4]  A. H. Murphy,et al.  A General Framework for Forecast Verification , 1987 .

[5]  M. Degroot,et al.  Optimal linear opinion pools , 1991 .

[6]  E. Langford,et al.  Is the Property of Being Positively Correlated Transitive? , 2001 .

[7]  Frank Lad,et al.  Learning from the Probability Assertions of Experts , 2003 .

[8]  Suk-Geun Hwang,et al.  Cauchy's Interlace Theorem for Eigenvalues of Hermitian Matrices , 2004, Am. Math. Mon..

[9]  Xin Guo,et al.  On the optimality of conditional expectation as a Bregman predictor , 2005, IEEE Trans. Inf. Theory.

[10]  Seung-Jean Kim,et al.  Maximum Likelihood Covariance Estimation with a Condition Number Constraint , 2006, 2006 Fortieth Asilomar Conference on Signals, Systems and Computers.

[11]  Robert L. Winkler,et al.  Probability Elicitation, Scoring Rules, and Competition Among Forecasters , 2007, Manag. Sci..

[12]  D. Moore,et al.  Use of absolute and comparative performance feedback in absolute and comparative judgments and decisions , 2008 .

[13]  S. Broomell,et al.  Why Are Experts Correlated? Decomposing Correlations Between Judges , 2009 .

[14]  Scott E. Page,et al.  Interpreted and generated signals , 2009, J. Econ. Theory.

[15]  David M. Pennock,et al.  Prediction without markets , 2010, EC '10.

[16]  Miguel Sousa Lobo,et al.  Human Judgement is Heavy Tailed: Empirical Evidence and Implications for the Aggregation of Estimates and Forecasts , 2010 .

[17]  T. Gneiting,et al.  Combining probability forecasts , 2010 .

[18]  Lyle H. Ungar,et al.  The Good Judgment Project: A Large Scale Test of Different Methods of Combining Expert Predictions , 2012, AAAI Fall Symposium: Machine Aggregation of Human Judgment.

[19]  Scott E. Page,et al.  Characterizing and aggregating agent estimates , 2013, AAMAS.

[20]  Mirai Tanaka,et al.  Positive definite matrix approximation with condition number constraint , 2014, Optim. Lett..

[21]  Lyle H. Ungar,et al.  Modeling Probability Forecasts via Information Diversity , 2014 .

[22]  Sydney E. Scott,et al.  Psychological Strategies for Winning a Geopolitical Forecasting Tournament , 2014, Psychological science.