Developing and testing a theoretical framework for assessing extended response questions in GCSE Science

This study aimed to develop, test, and validate a theoretical framework that could be used to write levels-based mark schemes for extended response questions in GCSE science. The work focused on questions which require students to give a scientific explanation or provide an argument. The development was informed by the work of researchers who have evaluated argumentation in the science classroom and also took into account the grade descriptors that are used to determine the cut scores for key grade boundaries during the awards process for GCSE science. The framework was used to write mark schemes for five questions from the January 2012 GCSE Science examinations. The mark schemes were used to mark scripts (n = 19 to 26) from those examinations. The marks awarded were compared with those given by the examiners who originally marked the questions. To ensure the theoretically-based mark scheme could be used by others, three senior examiners also marked two of the questions. Senior assessors (n = 12) for GCSE Science were asked, through an open response questionnaire, to comment on the framework and on its potential usefulness for writing levels-based mark schemes. Comparison of marks awarded using the theoretically-based scheme with those using the original scheme showed that the two schemes produced similar rank orders (Kendall’s coefficient τ = 0.61 to 0.83, n = 19 to 26). When other examiners used the theoretically-based scheme they awarded similar marks to those given by the researcher. These two outcomes suggest that the theoretically-based framework could be used for the proposed purpose. The senior assessors were generally positive about the usefulness of the scheme as a starting position for writing mark schemes and some recognised its potential to provide consistency of standards between different papers in the same session and across time.

[1]  Joseph Krajcik,et al.  Supporting Grade 5-8 Students in Constructing Explanations in Science: The Claim, Evidence, and Reasoning Framework for Talk and Writing , 2011 .

[2]  J. Biggs Assessment and Classroom Learning: a role for summative assessment? , 1998 .

[3]  G. Stobart Determining validity in national curriculum assessments , 2009 .

[4]  P. Newton The reliability of results from national curriculum testing in England , 2009 .

[5]  J. Osborne,et al.  Establishing the norms of scientific argumentation in classrooms , 2000 .

[6]  M. Jiménez-Aleixandre,et al.  Argument in High School Genetics. , 1997 .

[7]  S. Erduran,et al.  TAPping into argumentation: Developments in the application of Toulmin's Argument Pattern for studying science discourse , 2004 .

[8]  J. Osborne Science teaching methods: a rationale for practices , 2011 .

[9]  Sharon Price Bonham,et al.  Education and Employment , 1971 .

[10]  Developing and Applying Level Descriptors , 2003 .

[11]  Deborah Joy Corrigan,et al.  Valuing assessment in science education : pedagogy, curriculum, policy , 2013 .

[12]  R. Khishfe Explicit Nature of Science and Argumentation Instruction in the Context of Socioscientific Issues: An effect on student learning and transfer , 2014 .

[13]  Tom Bramley,et al.  The interrelations of features of questions, mark schemes and examinee responses and their impact upon marker agreement , 2011 .

[14]  Gregory J. Kelly,et al.  Students’ reasoning about electricity: combining performance assessments with argumentation analysis , 1998 .

[15]  Katherine L. McNeill,et al.  For whom is argument and explanation a necessary distinction? A response to Osborne and Patterson† , 2012 .

[16]  David A. Gillam,et al.  A Framework for K-12 Science Education: Practices, Crosscutting Concepts, and Core Ideas , 2012 .

[17]  Shirley Simon,et al.  Ideas, Evidence and Argument in Science , 2004 .

[18]  Paul Dalziel,et al.  Education and Skills , 2013 .

[19]  Helen R. Quinn,et al.  A Framework for K-12 Science Education: Practices, Crosscutting Concepts, and Core Ideas , 2013 .

[20]  Avi Hofstein,et al.  Peer Argumentation in the School Science Laboratory—Exploring effects of task features , 2011 .

[21]  Paul Black,et al.  Testing: Friend or Foe?: Theory and Practice of Assessment and Testing , 1997 .

[22]  J. Osborne,et al.  Supporting and Promoting Argumentation Discourse in Science Education , 2002 .

[23]  Sibel Erduran,et al.  Argumentation in Science Education: Perspectives from Classroom-Based Research , 2007 .

[24]  Colin Robson,et al.  Real world research : a resource for users of social research methods in applied settings , 2011 .

[25]  J. Osborne,et al.  The place of argumentation in the pedagogy of school science , 1999 .

[26]  Shirley Simon,et al.  Enhancing the quality of argumentation in school science , 2004 .

[27]  Robin Millar,et al.  Twenty First Century Science: Insights from the Design and Implementation of a Scientific Literacy Approach in School Science , 2006 .

[28]  Ying-Chih Chen,et al.  Examining the Integration of Talk and Writing for Student Knowledge Construction through Argumentation. , 2011 .

[29]  W. Doyle,et al.  The practicality ethic in teacher decision-making , 1977 .

[30]  J. Baird,et al.  Research evidence relating to proposals for reform of the GCSE , 2013 .

[31]  Irenka Suto,et al.  Who should mark what? A study of factors affecting marking accuracy in a biology examination , 2011 .

[32]  Shirley Simon,et al.  Argumentation in School Science: Breaking the Tradition of Authoritative Exposition Through a Pedagogy that Promotes Discussion and Reasoning , 2009 .

[33]  Margaret Spear,et al.  The influence of contrast effects upon teachers’ marks , 1997 .

[34]  Belita Gordon,et al.  Resolving Score Differences in the Rating of Writing Samples: Does Discussion Improve the Accuracy of Scores? , 2005 .

[35]  Jonathan Osborne,et al.  Language and Literacy in Science Education , 2001 .

[36]  Laurence Simonneaux,et al.  Argumentation in Science Education: An Overview , 2007 .

[37]  S. Chamberlain Do marking reliability studies have validity , 2008 .

[39]  M. Fullan The New Meaning of Educational Change , 1990 .

[40]  E. Wenger Communities of Practice and Social Learning Systems , 2000 .

[41]  Bernice W. Polemis Nonparametric Statistics for the Behavioral Sciences , 1959 .

[42]  Douglas Walton,et al.  Argument Structure: A Pragmatic Theory , 1996 .

[43]  What determines GCSE marking accuracy? An exploration of expertise among maths and physics markers , 2008 .

[44]  S. Toulmin The uses of argument , 1960 .

[45]  J. Andrews,et al.  British Educational Research Association, University of Warwick , 2006 .

[46]  Jackie Greatorex,et al.  What makes marking reliable? Experiments with UK examinations , 2004 .

[47]  Mike Watts,et al.  Science in the national curriculum , 1991 .

[48]  Jonathan Osborne,et al.  Scientific argument and explanation: A necessary distinction? , 2011 .

[49]  J. Osborne,et al.  Authors' response to “For whom is argument and explanation a necessary distinction? A response to Osborne and Patterson” by Berland and McNeill , 2012 .

[50]  Michelle Meadows,et al.  A review of the literature on marking reliability , 2005 .

[51]  M. Frisch,et al.  Authors' response to: Mortality rates for same-sex married individuals compared with opposite-sex married individuals. , 2015, International journal of epidemiology.

[52]  R. Duschl,et al.  "Doing the Lesson" or "Doing Science": Argument in High School Genetics , 2000 .

[53]  Robin Millar,et al.  Improving Science Education: Why Assessment Matters , 2013 .

[54]  Andy P. Field,et al.  Discovering Statistics Using SPSS , 2000 .