Reasoning up and down a Food Chain: Using an Assessment Framework to Investigate Students' Middle Knowledge.

Being able to make claims about what students know and can do in science involves gathering systematic evidence of students' knowledge and abilities. This paper describes an assessment system designed to elicit information from students at many placements along developmental trajectories and demonstrates how this system was used to gather principled evidence of how students reason about food web and food chain disturbances. Specifically, this assessment system was designed to gather information about students' intermediary or middle knowledge on a pathway toward more sophisticated understanding. Findings indicate that in association with a curricular intervention, student gains were significant. However, despite overall gains, some students still struggled to explain what might happen during a disturbance to an ecosystem. In addition, this paper discusses the importance of having a cognitive framework to guide task design and interpretation of evidence. This framework allowed for the gathering of detailed information, which provided insights into the intricacies of how students reason about scientific scenarios. In particular, this assessment system allowed for the illustration of multiple types of middle knowledge that students may possess, indicating that there are multiple “messy middles” students may move through as they develop the ability to reason about complex scientific situations. © 2009 Wiley Periodicals, Inc. Sci Ed94:259–281, 2010

[1]  K. A. Ericsson,et al.  Protocol Analysis: Verbal Reports as Data , 1984 .

[2]  M. Linn,et al.  Scientific arguments as learning artifacts: designing for learning from the web with KIE , 2000 .

[3]  C. Fox,et al.  Applying the Rasch Model: Fundamental Measurement in the Human Sciences , 2001 .

[4]  D. Kuhn Children and adults as intuitive scientists. , 1989, Psychological review.

[5]  Miriam Reiner,et al.  Conceptual classroom environment - a system view of learning , 2001 .

[6]  Charles R. Barman,et al.  An Analysis of High School Students' Concepts & Textbook Presentations of Food Chains & Food Webs. , 1994 .

[7]  Los Angeles,et al.  Evaluation.; Center for Research on Evaluation, Standards, and Student Testing, , 1998 .

[8]  H. Schweingruber,et al.  TAKING SCIENCE TO SCHOOL: LEARNING AND TEACHING SCIENCE IN GRADES K-8 , 2007 .

[9]  Raymond J. Adams,et al.  Rasch models for item bundles , 1995 .

[10]  Amelia Wenk Gotwals,et al.  Measuring Students' Scientific Content and Inquiry Reasoning , 2006, ICLS.

[11]  Alija Kulenović,et al.  Standards for Educational and Psychological Testing , 1999 .

[12]  R. Glaser,et al.  Knowing What Students Know: The Science and Design of Educational Assessment , 2001 .

[13]  Meryl W. Bertenthal,et al.  Systems for state science assessment , 2005 .

[14]  Amelia Wenk Gotwals The nature of students' science knowledge base: Using assessment to paint a picture , 2006 .

[15]  J. Osborne,et al.  Establishing the norms of scientific argumentation in classrooms , 2000 .

[16]  Hee-Sun Lee Scaffolding elementary students' authentic inquiry through a written science curriculum. , 2003 .

[17]  Joseph Krajcik,et al.  Supporting Students' Construction of Scientific Explanations by Fading Scaffolds in Instructional Materials , 2006 .

[18]  John Leach,et al.  Children's ideas about ecology 3: ideas found in children aged 5‐16 about the interdependency of organisms , 1996 .

[19]  Georg Rasch,et al.  Probabilistic Models for Some Intelligence and Attainment Tests , 1981, The SAGE Encyclopedia of Research Design.

[20]  R. Sawyer The Cambridge Handbook of the Learning Sciences: Introduction , 2014 .

[21]  R. Almond,et al.  Focus Article: On the Structure of Educational Assessments , 2003 .

[22]  Raymond J. Adams,et al.  The Multidimensional Random Coefficients Multinomial Logit Model , 1997 .

[23]  S. K. Alexander Food web analysis: an ecosystem approach , 1982 .

[24]  P. Black,et al.  Classroom Assessment and the National Science Education Standards. , 2001 .

[25]  S. Toulmin The uses of argument , 1960 .

[26]  Nancy Butler Songer,et al.  The Cambridge Handbook of the Learning Sciences: BioKIDS , 2005 .

[27]  Hee-Sun Lee,et al.  Making authentic science accessible to students , 2003 .

[28]  M. Wilson Detecting and Interpreting Local Item Dependence Using a Fannily of Rasch Models , 1988 .

[29]  Robert J. Mislevy,et al.  Psychometric Principles in Student Assessment , 2003 .

[30]  Ben Kelcey,et al.  How and when does complex reasoning occur? Empirically driven development of a learning progression focused on complex reasoning about biodiversity , 2009 .

[31]  William A. Sandoval,et al.  Conceptual and Epistemic Aspects of Students' Scientific Explanations , 2003 .

[32]  Kathleen Hogan,et al.  Assessing students' systems reasoning in ecology , 2000 .

[33]  James W Pellegrino,et al.  Technology and Testing , 2009, Science.

[34]  S. Messick The Interplay of Evidence and Consequences in the Validation of Performance Assessments , 1994 .

[35]  Brian J. Reiser,et al.  Scaffolding Complex Learning: The Mechanisms of Structuring and Problematizing Student Work , 2004, The Journal of the Learning Sciences.

[36]  Karen Draney,et al.  Objective measurement : theory into practice , 1992 .

[37]  Fernando Flores,et al.  Preconceptions and relations used by children in the construction of food chains , 1994 .