The Demographic and Political Composition of Mechanical Turk Samples

One of the most notable recent developments in survey research is the increased usage of online convenience samples drawn from Amazon’s Mechanical Turk (MTurk). While scholars have noted various social and political differences (e.g., age, partisanship) between MTurk and population-based samples, the breadth and depth of these variations remain unclear. We investigate the extent to which MTurk samples differ from population samples, and the underlying nature of these differences. We do so by replicating items from the population-based American National Election Studies (ANES) 2012 Time Series Study in a survey administered to a sample of MTurk respondents. With few exceptions, we not only find that MTurk respondents differ significantly from respondents completing the 2012 ANES via the Web but also that most differences are reduced considerably when controlling for easily measurable sample features. Thus, MTurk respondents do not appear to differ fundamentally from population-based respondents in unmeasurable ways. This suggests that MTurk data can be used to advance research programs, particularly if researchers measure and account for a range of political and demographic variables as needed.

[1]  Gabriel S. Lenz,et al.  Sources of Bias in Retrospective Decision Making: Experimental Evidence on Voters’ Limitations in Controlling Incumbents , 2012, American Political Science Review.

[2]  Adam Seth Levine,et al.  Cross-Sample Comparisons and External Validity , 2014, Journal of Experimental Political Science.

[3]  Mario Callegaro,et al.  Online panel research: History, concepts, applications and a look at the future , 2014 .

[4]  Yichen Wang,et al.  Data collection mode effect on feeling thermometer questions: A comparison of face-to-face and Web surveys , 2015, Comput. Hum. Behav..

[5]  M. Couper A REVIEW OF ISSUES AND APPROACHES , 2000 .

[6]  Bella Struminskaya,et al.  Assessing representativeness of a probability‐based online panel in Germany , 2014 .

[7]  Mick P. Couper,et al.  The Future of Modes of Data Collection , 2011 .

[8]  G. Huber,et al.  Assessing the Programmatic Equivalence Assumption in Question Wording Experiments Understanding Why Americans Like Assistance to the Poor More Than Welfare , 2013 .

[9]  Thomas J. Leeper,et al.  The Generalizability of Survey Experiments* , 2015, Journal of Experimental Political Science.

[10]  Panagiotis G. Ipeirotis Demographics of Mechanical Turk , 2010 .

[11]  Duncan J. Watts,et al.  Financial incentives and the "performance of crowds" , 2009, HCOMP '09.

[12]  D. Tingley,et al.  “Who are these people?” Evaluating the demographic characteristics and political preferences of MTurk survey respondents , 2015 .

[13]  Panagiotis G. Ipeirotis,et al.  Running Experiments on Amazon Mechanical Turk , 2010, Judgment and Decision Making.

[14]  Jon A. Krosnick,et al.  Research Synthesis AAPOR Report on Online Panels , 2010 .

[15]  Krista Casler,et al.  Separate but equal? A comparison of participants and data gathered via Amazon's MTurk, social media, and face-to-face behavioral testing , 2013, Comput. Hum. Behav..

[16]  Jesse Chandler,et al.  Selectively Recruiting Participants from Amazon Mechanical Turk Using Qualtrics , 2012 .

[17]  Diana C. Mutz Population-Based Survey Experiments , 2011 .

[18]  Adam J. Berinsky,et al.  Evaluating Online Labor Markets for Experimental Research: Amazon.com's Mechanical Turk , 2012, Political Analysis.

[19]  Kevin Arceneaux Cognitive Biases and the Strength of Political Arguments , 2012 .

[20]  D. Green,et al.  Modeling Heterogeneous Treatment Effects in Survey Experiments with Bayesian Additive Regression Trees , 2012 .