Principled Assessment of Student Learning in High School Computer Science

As K-12 computer science (CS) initiatives scale throughout the U.S., educators face increasing pressure from their school systems to provide evidence about student learning on hard-to-measure CS outcomes. At the same time, researchers studying curriculum implementation and student learning want reliable measures of how students apply their CS knowledge. This paper describes a two-year validation study focused on end-of-unit and cumulative assessments for Exploring Computer Science, an introductory high school CS curriculum. To develop the assessments, we applied a principled methodology called Evidence-Centered Design (ECD) to (1) work with various stakeholders to identify the important computer science skills to measure, (2) map those skills to a model of evidence that can support inferences about those skills, and (3) develop assessment tasks that elicit that evidence. Using ECD, we created assessments that measure the practices of computational thinking, in contrast to assessments that only measure CS conceptual knowledge. We iteratively developed and piloted the assessments with 941 students over two years and collected three types of validity evidence based on contemporary psychometric standards: test content, internal structure, and student response processes. Results show that reliability was moderate to high for each of the unit assessments; the assessment tasks within each assessment are well aligned with each other and with the targeted learning goals; and average scores were in the 60 to 70 percent range. These results indicate that the assessments validly measure students' computational thinking practices covered in the introductory CS curriculum. We discuss the broader issues we faced of balancing the need to use the assessment results for evaluation and research, and demands from teachers for use in the classroom.

[1]  Andrea C. Arpaci-Dusseau,et al.  Computer science principles: analysis of a proposed advanced placement course , 2013, SIGCSE '13.

[2]  E. McColl Cognitive Interviewing. A Tool for Improving Questionnaire Design , 2006, Quality of Life Research.

[3]  Mark Guzdial,et al.  A multi-national, multi-institutional study of assessment of programming skills of first-year CS students , 2001, ITiCSE-WGR '01.

[4]  Jill Denner,et al.  The fairy performance assessment: measuring computational thinking in middle school , 2012, SIGCSE '12.

[5]  Joanna S. Gorin Test Design with Cognition in Mind , 2007 .

[6]  Mark Wilson,et al.  Constructing Measures: An Item Response Modeling Approach , 2004 .

[7]  K. A. Ericsson,et al.  Protocol Analysis: Verbal Reports as Data , 1984 .

[8]  Catharine Brand Lowering Barriers to Interaction: Programming without Code , 2007 .

[9]  Jill Denner,et al.  Children Programming Games: A Strategy for Measuring Computational Learning , 2015, TOCE.

[10]  Educational Evaluation Standards for Educational and Psychological Testing , 1999 .

[11]  Steven M. Downing,et al.  Handbook of test development , 2006 .

[12]  Alexander Repenning,et al.  Towards the Automatic Recognition of Computational Thinking for Adaptive Visual Language Learning , 2010, 2010 IEEE Symposium on Visual Languages and Human-Centric Computing.

[13]  Robert J. Mislevy,et al.  Implications of Evidence‐Centered Design for Educational Testing , 2007 .

[14]  Mark Guzdial,et al.  Developing a validated assessment of fundamental CS1 concepts , 2010, SIGCSE.

[15]  R. Linn Educational measurement, 3rd ed. , 1989 .

[16]  Shuchi Grover,et al.  Computational Thinking in K–12 , 2013 .

[17]  Michael C. Loui,et al.  Creating the digital logic concept inventory , 2010, SIGCSE.

[18]  Maarten van Someren,et al.  The Think Aloud Method: A Practical Guide to Modelling Cognitive Processes , 1994 .

[19]  Mitchel Resnick,et al.  Programming by choice: urban youth learning programming with scratch , 2008, SIGCSE '08.

[20]  Mark Guzdial,et al.  Introductory Computing Construct Use in an End-User Programming Community , 2007, IEEE Symposium on Visual Languages and Human-Centric Computing (VL/HCC 2007).

[21]  Bazil Pârv,et al.  A nationwide exam as a tool for improving a new curriculum , 2013, ITiCSE '13.

[22]  Russell G. Almond,et al.  On the Structure of Educational Assessments, CSE Technical Report. , 2003 .

[23]  Gautam Biswas,et al.  Integrating computational thinking with K-12 science education using agent-based computation: A theoretical framework , 2013, Education and Information Technologies.

[24]  Fred Martin,et al.  Integrating computational thinking across the K--8 curriculum , 2014, Inroads.

[25]  David C. Webb,et al.  Troubleshooting assessment: an authentic problem solving activity for it education , 2010 .

[26]  Chris Stephenson,et al.  Exploring the Science Framework and NGSS: Computational Thinking in the Science Classroom , 2014 .

[27]  Sibel Erduran Developing assessments for the next generation science standards , 2017 .

[28]  Michael S. Horn,et al.  Fostering computational literacy in science classrooms , 2014, CACM.

[29]  Lorin W. Anderson,et al.  OBJECTIVES, EVALUATION, AND THE IMPROVEMENT OF EDUCATION , 2005 .

[30]  Linda M. Seiter,et al.  Modeling the learning progressions of computational thinking of primary grade students , 2013, ICER.

[31]  Zachary Dodds,et al.  MyCS at 5: Assessing a Middle-years CS Curriculum , 2016, SIGCSE.

[32]  Joanna Goode,et al.  Beyond curriculum: the exploring computer science program , 2012, INROADS.

[33]  Jeannette M. Wing Computational thinking and thinking about computing , 2008, Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences.

[34]  Chris Stephenson,et al.  Running on Empty: the Failure to Teach K--12 Computer Science in the Digital Age , 2010 .