Generalizing from Survey Experiments Conducted on Mechanical Turk: A Replication Approach

To what extent do survey experimental treatment effect estimates generalize to other populations and contexts? Survey experiments conducted on convenience samples have often been criticized on the grounds that subjects are sufficiently different from the public at large to render the results of such experiments uninformative more broadly. In the presence of moderate treatment effect heterogeneity, however, such concerns may be allayed. I provide evidence from a series of 15 replication experiments that results derived from convenience samples like Amazon’s Mechanical Turk are similar to those obtained from national samples. Either the treatments deployed in these experiments cause similar responses for many subject types or convenience and national samples do not differ much with respect to treatment effect moderators. Using evidence of limited within-experiment heterogeneity, I show that the former is likely to be the case. Despite a wide diversity of background characteristics across samples, the effects uncovered in these experiments appear to be relatively homogeneous.

[1]  C. B. Colby The weirdest people in the world , 1973 .

[2]  S. Holm A Simple Sequentially Rejective Multiple Test Procedure , 1979 .

[3]  L. Cronbach,et al.  Designing evaluations of educational and social programs , 1983 .

[4]  D. O. Sears College sophomores in the laboratory: Influences of a narrow data base on social psychology's view of human nature. , 1986 .

[5]  H. White,et al.  A Reality Check for Data Snooping , 2000 .

[6]  Mark R. Joslyn,et al.  Gun Policy, Opinion, Tragedy, and Blame Attribution: The Conditional Influence of Issue Frames , 2001, The Journal of Politics.

[7]  R. McDermott Experimental Methodology in Political Science , 2002, Political Analysis.

[8]  J. Lucas,et al.  Theory-Testing, Generalization, and the Problem of External Validity* , 2003 .

[9]  Ted Brader,et al.  Striking a Responsive Chord: How Political Ads Motivate and Persuade Voters by Appealing to Emotions , 2005 .

[10]  J. Ioannidis Why Most Published Research Findings Are False , 2005, PLoS medicine.

[11]  M. Hiscox,et al.  Through a Glass and Darkly: Attitudes Toward International Trade and the Curious Effects of Issue Framing , 2006, International Organization.

[12]  M. Peffley,et al.  Persuasion and Resistance: Race and the Death Penalty in America , 2007 .

[13]  Steven D. Levitt,et al.  What Do Laboratory Experiments Measuring Social Preferences Reveal About the Real World , 2007 .

[14]  J. Transue,et al.  Identity Salience, Identity Acceptance, and Racial Policy Attitudes: American National Identity as a Uniting Force , 2007 .

[15]  S. Levinson,et al.  WEIRD languages have misled us, too , 2010, Behavioral and Brain Sciences.

[16]  A. Gerber,et al.  Publication Bias in Two Political Behavior Literatures , 2010 .

[17]  James N. Druckman,et al.  Dynamic Public Opinion: Communication Effects over Time , 2010, American Political Science Review.

[18]  Stephen P. Nicholson Polarizing Cues: POLARIZING CUES , 2012 .

[19]  Donald P. Green,et al.  Field Experiments: Design, Analysis, and Interpretation , 2012 .

[20]  Peter M. Aronow,et al.  On equivalencies between design-based and regression-based variance estimators for randomized experiments , 2012 .

[21]  Stephen P. Nicholson,et al.  Polarizing Cues , 2011 .

[22]  Adam J. Berinsky,et al.  Evaluating Online Labor Markets for Experimental Research: Amazon.com's Mechanical Turk , 2012, Political Analysis.

[23]  Elizabeth Tipton Improving Generalizations From Experiments Using Propensity Score Subclassification , 2013 .

[24]  Daniel W Webster,et al.  Effects of news media messages about mass shootings on attitudes toward persons with serious mental illness and public support for gun control policies. , 2013, The American journal of psychiatry.

[25]  Macartan Humphreys,et al.  Fishing, Commitment, and Communication: A Proposal for Comprehensive Nonbinding Research Registration , 2012, Political Analysis.

[26]  D. Green,et al.  Assessing the Correspondence between Experimental Results Obtained in the Lab and Field: A Review of Recent Social Science Research* , 2014, Political Science Research and Methods.

[27]  Kate A. Ratliff,et al.  Non-Naïve Participants Can Reduce Effect Sizes , 2014 .

[28]  Peng Ding,et al.  Randomization inference for treatment effect variation , 2014, 1412.5000.

[29]  Neil Malhotra,et al.  Publication bias in the social sciences: Unlocking the file drawer , 2014, Science.

[30]  Jennifer A. Richeson,et al.  More Diverse Yet Less Tolerant? How the Increasingly Diverse Racial Landscape Affects White Americans’ Racial Attitudes , 2014, Personality & social psychology bulletin.

[31]  Adam Seth Levine,et al.  Cross-Sample Comparisons and External Validity , 2014, Journal of Experimental Political Science.

[32]  Leif D. Nelson,et al.  P-Curve: A Key to the File Drawer , 2013, Journal of experimental psychology. General.

[33]  Thomas J. Leeper,et al.  The Generalizability of Survey Experiments* , 2015, Journal of Experimental Political Science.

[34]  D. Tingley,et al.  “Who are these people?” Evaluating the demographic characteristics and political preferences of MTurk survey respondents , 2015 .

[35]  Ben R. Newell,et al.  The average laboratory samples a population of 7,300 Amazon Mechanical Turk workers , 2015, Judgment and Decision Making.

[36]  Michael C. Frank,et al.  Estimating the reproducibility of psychological science , 2015, Science.

[37]  Kate A. Ratliff,et al.  Using Nonnaive Participants Can Reduce Effect Sizes , 2015, Psychological science.

[38]  Timothy D. Wilson,et al.  Comment on “Estimating the reproducibility of psychological science” , 2016, Science.

[39]  Neil Malhotra,et al.  Does Media Coverage of Partisan Polarization Affect Political Attitudes? , 2016 .

[40]  Economists and Public Opinion: Expert Consensus and Economic Policy Judgments , 2016, The Journal of Politics.

[41]  Jonathan Mummolo,et al.  Assessing the Breadth of Framing Effects , 2017 .