Self-coding: A method to assess semantic validity and bias when coding open-ended responses

Open-ended survey questions can provide researchers with nuanced and rich data, but content analysis is subject to misinterpretation and can introduce bias into subsequent analysis. We present a simple method to improve the semantic validity of a codebook and test for bias: a “self-coding” method where respondents first provide open-ended responses and then self-code those responses into categories. We demonstrated this method by comparing respondents’ self-coding to researcher-based coding using an established codebook. Our analysis showed significant disagreement between the codebook’s assigned categorizations of responses and respondents’ self-codes. Moreover, this technique uncovered instances where researcher-based coding disproportionately misrepresented the views of certain demographic groups. We propose using the self-coding method to iteratively improve codebooks, identify bad-faith respondents, and, perhaps, to replace researcher-based content analysis.

[1]  James N. Druckman,et al.  The Decline of the Death Penalty and the Discovery of Innocence , 2008 .

[2]  Roei Davidson,et al.  Who's to Blame When a Business Fails? How Journalistic Death Metaphors Influence Responsibility Attributions , 2011 .

[3]  Arthur Lupia,et al.  Cambridge Handbook of Experimental Political Science: Political Knowledge , 2011 .

[4]  Marijke Breuning,et al.  Rationalization and Derivation Processes in Survey Studies of Political Candidate Evaluation , 1994 .

[5]  K. Crittenden,et al.  Improving Coding Reliability for Open-Ended Questions , 1977 .

[6]  Gregory A. Caldeira,et al.  Knowing the Supreme Court? A Reconsideration of Public Ignorance of the High Court , 2009, The Journal of Politics.

[7]  Christopher Wlezien,et al.  On the salience of political issues: The problem with ‘most important problem’ ☆ , 2005 .

[8]  David G. Rand,et al.  Structural Topic Models for Open‐Ended Survey Responses , 2014, American Journal of Political Science.

[9]  Jane Green,et al.  Owning the issue agenda: Party strategies and vote choices in British elections , 2008 .

[10]  G. F. Bishop,et al.  EXPERIMENTS WITH THE MIDDLE RESPONSE ALTERNATIVE IN SURVEY QUESTIONS , 1987 .

[11]  Jolene D. Smyth,et al.  Open-Ended Questions in Web Surveys Can Increasing the Size of Answer Boxes and Providing Extra Verbal Instructions Improve Response Quality? , 2009 .

[12]  Matthew O. Hunt African American, Hispanic, and White Beliefs about Black/White Inequality, 1977-2004 , 2007 .

[13]  Amber E. Boydstun,et al.  Framing, identity, and responsibility: do episodic vs. thematic framing effects vary by target population? , 2019, Politics, Groups, and Identities.

[14]  Gary Hickey,et al.  A multi-stage approach to the coding of data from open-ended questions. , 1996, Nurse researcher.

[15]  Lynn Vavreck,et al.  The Mass Media and the Public's Assessments of Presidential Candidates, 1952–2000 , 2007, The Journal of Politics.

[16]  Deen Freelon,et al.  ReCal OIR : Ordinal , Interval , and Ratio Intercoder Reliability as a Web Service , 2013 .

[17]  S. Presser,et al.  Questions and Answers in Attitude Surveys: Experiments on Question Form, Wording, and Context , 1996 .

[18]  S. Feldman,et al.  A Simple Theory of the Survey Response: Answering Questions versus Revealing Preferences , 1992 .

[19]  J. Krosnick,et al.  AN EVALUATION OF A COGNITIVE THEORY OF RESPONSE-ORDER EFFECTS IN SURVEY MEASUREMENT , 1987 .

[20]  Cheryl Boudreau,et al.  What is Political Knowledge? , 1961, The Review of Politics.

[21]  Thomas J. Leeper,et al.  The Generalizability of Survey Experiments* , 2015, Journal of Experimental Political Science.

[22]  H. Schuman,et al.  Problems in the Use of Survey Questions to Measure Public Opinion , 1987, Science.

[23]  A. Coppock Generalizing from Survey Experiments Conducted on Mechanical Turk: A Replication Approach , 2018, Political Science Research and Methods.

[24]  Mark Swyngedouw The subjective cognitive and affective map of extreme right voters: using open-ended questions in exit polls , 2001 .

[25]  A. Eagly,et al.  Gender gaps in sociopolitical attitudes: a social psychological analysis. , 2004, Journal of personality and social psychology.

[26]  Adam J. Berinsky,et al.  Evaluating Online Labor Markets for Experimental Research: Amazon.com's Mechanical Turk , 2012, Political Analysis.

[27]  Dirk Heerwegh,et al.  Mode Differences Between Face-to-Face and Web Surveys: An Experimental Investigation of Data Quality and Social Desirability Effects , 2009 .

[28]  John G. Geer,et al.  WHAT DO OPEN-ENDED QUESTIONS MEASURE? , 1988 .

[29]  Mark R. Joslyn,et al.  Gun Policy, Opinion, Tragedy, and Blame Attribution: The Conditional Influence of Issue Frames , 2001, The Journal of Politics.

[30]  Vasja Vehovar,et al.  Open-ended vs. close-ended questions in Web questionnaires , 2003 .

[31]  A. Acquisti,et al.  Reputation as a sufficient condition for data quality on Amazon Mechanical Turk , 2013, Behavior Research Methods.

[32]  Michael A. Xenos,et al.  Media Framing and Effective Public Deliberation , 2000 .

[33]  P. Lazarsfeld THE CONTROVERSY OVER DETAILED INTERVIEWS—AN OFFER FOR NEGOTIATION , 1944 .

[34]  The shape of and solutions to the MTurk quality crisis , 2020, Political Science Research and Methods.

[35]  Norbert Schwarz,et al.  A Cognitive Model of Response-Order Effects in Survey Measurement , 1992 .

[36]  Kevin Arceneaux The Conditional Impact of Blame Attribution on the Relationship Between Economic Adversity and Turnout , 2003 .

[37]  S. Iyengar Is anyone responsible? How television frames political issues. , 1991 .

[38]  Ivar Krumpal Determinants of social desirability bias in sensitive surveys: a literature review , 2013 .