Objective Social Choice: Using Auxiliary Information to Improve Voting Outcomes

How should one combine noisy information from diverse sources to make an inference about an objective ground truth? This frequently recurring, normative question lies at the core of statistics, machine learning, policy-making, and everyday life. It has been called "combining forecasts", "meta-analysis", "ensembling", and the "MLE approach to voting", among other names. Past studies typically assume that noisy votes are identically and independently distributed (i.i.d.), but this assumption is often unrealistic. Instead, we assume that votes are independent but not necessarily identically distributed and that our ensembling algorithm has access to certain auxiliary information related to the underlying model governing the noise in each vote. In our present work, we: (1) define our problem and argue that it reflects common and socially relevant real world scenarios, (2) propose a multi-arm bandit noise model and count-based auxiliary information set, (3) derive maximum likelihood aggregation rules for ranked and cardinal votes under our noise model, (4) propose, alternatively, to learn an aggregation rule using an order-invariant neural network, and (5) empirically compare our rules to common voting rules and naive experience-weighted modifications. We find that our rules successfully use auxiliary information to outperform the naive baselines.

[1]  J. Fleiss Review papers : The statistical basis of meta-analysis , 1993 .

[2]  Vincent Conitzer,et al.  Handbook of Computational Social Choice , 2016 .

[3]  Ariel D. Procaccia,et al.  When do noisy votes reveal the truth? , 2013, EC '13.

[4]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[5]  C. Granger,et al.  Improved methods of combining forecasts , 1984 .

[6]  J. Dickinson Some Statistical Results in the Combination of Forecasts , 1973 .

[7]  Lior Rokach,et al.  Ensemble-based classifiers , 2010, Artificial Intelligence Review.

[8]  C. L. Mallows NON-NULL RANKING MODELS. I , 1957 .

[9]  John A. Weymark,et al.  Interpersonal Comparisons of Well-being: A reconsideration of the Harsanyi–Sen debate on utilitarianism , 1991 .

[10]  Vincent Conitzer The maximum likelihood approach to voting on social networks , 2013, 2013 51st Annual Allerton Conference on Communication, Control, and Computing (Allerton).

[11]  Paul A. Gompers,et al.  CORPORATE GOVERNANCE AND EQUITY PRICES , 2002 .

[12]  L. Cooper,et al.  When Networks Disagree: Ensemble Methods for Hybrid Neural Networks , 1992 .

[13]  Ariel D. Procaccia,et al.  The Distortion of Cardinal Preferences in Voting , 2006, CIA.

[14]  J. M. Bates,et al.  The Combination of Forecasts , 1969 .

[15]  Bhavana Dalvi,et al.  A Dataset of Peer Reviews (PeerRead): Collection, Insights and NLP Applications , 2018, NAACL.

[16]  Richard Y. Chen,et al.  UCB EXPLORATION VIA Q-ENSEMBLES , 2018 .

[17]  C. Granger Invited review combining forecasts—twenty years later , 1989 .

[18]  Jürgen Schmidhuber,et al.  Long Short-Term Memory , 1997, Neural Computation.

[19]  J. Harsanyi Cardinal Welfare, Individualistic Ethics, and Interpersonal Comparisons of Utility , 1955 .

[20]  A. Sen,et al.  Collective Choice and Social Welfare , 2017 .

[21]  H. Young Condorcet's Theory of Voting , 1988, American Political Science Review.

[22]  Thorsten Joachims,et al.  The K-armed Dueling Bandits Problem , 2012, COLT.

[23]  R. Clemen Combining forecasts: A review and annotated bibliography , 1989 .

[24]  Shane Legg,et al.  Deep Reinforcement Learning from Human Preferences , 2017, NIPS.

[25]  Alexander J. Smola,et al.  Deep Sets , 2017, 1703.06114.

[26]  Christian Genest,et al.  Combining Probability Distributions: A Critique and an Annotated Bibliography , 1986 .

[27]  Lu Hong,et al.  Groups of diverse problem solvers can outperform groups of high-ability problem solvers. , 2004, Proceedings of the National Academy of Sciences of the United States of America.

[28]  K. McConway Marginalization and Linear Opinion Pools , 1981 .

[29]  J. Dickinson Some Comments on the Combination of Forecasts , 1975 .

[30]  Ariel D. Procaccia,et al.  Ranked Voting on Social Networks , 2015, IJCAI.

[31]  Kenneth F. Wallis,et al.  Combining forecasts – forty years later , 2011 .

[32]  David D. Lewis,et al.  Naive (Bayes) at Forty: The Independence Assumption in Information Retrieval , 1998, ECML.

[33]  Ioannis Caragiannis,et al.  Learning a Ground Truth Ranking Using Noisy Approval Votes , 2017, IJCAI.

[34]  Vincent Conitzer,et al.  Common Voting Rules as Maximum Likelihood Estimators , 2005, UAI.

[35]  Robert A. Jacobs,et al.  Methods For Combining Experts' Probability Assessments , 1995, Neural Computation.

[36]  Kenneth O. May,et al.  A Set of Independent Necessary and Sufficient Conditions for Simple Majority Decision , 1952 .

[37]  Christian Genest,et al.  Allocating the weights in the linear opinion pool , 1990 .

[38]  Volker Tresp,et al.  Combining Estimators Using Non-Constant Weighting Functions , 1994, NIPS.

[39]  Thomas G. Dietterich Multiple Classifier Systems , 2000, Lecture Notes in Computer Science.

[40]  R. J. Williams,et al.  Simple Statistical Gradient-Following Algorithms for Connectionist Reinforcement Learning , 2004, Machine Learning.