Effect of Examinee Certainty on Probabilistic Test Scores and a Comparison of Scoring Methods for Probabilistic Responses.

Abstract : This study was an attempt to alleviate some of the difficulties inherent in multiple-choice items by having examinees respond to multiple-choice items in a probabilistic manner. Using this format, examinees are able to respond to each alternative and to provide indications of any partial knowledge they may possess concerning the item. The items used in this study were 30 multiple-choice analogy items. Each item was scored using five different scoring formulas. Three of these scoring formulas were reproducing scoring systems. The fourth scoring method used the probability assigned to the correct alternative as the item score, and the fifth used a function of the absolute difference between the correct response vector for the four alternatives and the actual points assigned to each alternative as the item score. Total test scores for all of the scoring methods were obtained by summing individual item scores. Several studies using probabilistic response methods have shown the effect of a response-style variable, called certainty or risk taking, on scores obtained from probabilistic responses. Results from this study showed a small effect of certainty on the probabilistic scores in terms of the validity of the scores but no effect at all on the factor structure or internal consistency of the scores. The reproducing scoring systems may have an advantage because they maximize examinee's scores when examinees respond honestly, while honest responses will not necessarily maximize an examinee's score with the other two methods. If a reproducing scoring systems is used for this reason, the spherical scoring formula is recommended, since it was the most internally consistent and showed the str ongest first factor of the reproducing scoring systems.

[1]  Paul Horst,et al.  Obtaining a composite measure from a number of different measures of the same attribute , 1936 .

[2]  S. S. Wilks Weighting systems for linear functions of correlated variables when there is no dependent variable , 1938 .

[3]  H. Gulliksen Theory of mental tests , 1952 .

[4]  John Schmid,et al.  Some Modifications of the Multiple-Choice Item , 1953 .

[5]  Palmer O. Johnson,et al.  Modern statistical methods : descriptive and inductive , 1959 .

[6]  B. deFinetti,et al.  METHODS FOR DISCRIMINATING LEVELS OF PARTIAL KNOWLEDGE CONCERNING A TEST ITEM. , 1965, The British journal of mathematical and statistical psychology.

[7]  E. H. Shuford,et al.  Admissible probability measurement procedures , 1966, Psychometrika.

[8]  P M Bentler,et al.  Alpha-maximized factor analysis (alphamax): Its relation to alpha and canonical factor analysis , 1968, Psychometrika.

[9]  Robert M. Rippey A COMPARISON OF FIVE DIFFERENT SCORING FUNCTIONS FOR CONFIDENCE TESTS1 , 1970 .

[10]  L. Aiken Scoring for Partial Knowledge on the Generalized Rearrangement Item , 1970 .

[11]  S. Jacobs Correlates of Unwarranted Confidence in Responses to Objective Test Items. , 1971 .

[12]  R. Koehler A COMPARISON OF THE VALIDITIES OF CONVENTIONAL CHOICE TESTING AND VARIOUS CONFIDENCE MARKING PROCEDURES , 1971 .

[13]  Robert F. Boldt,et al.  PERSONALITY INFLUENCES ON CONFIDENCE TEST SCORES , 1972 .

[14]  D. Gilman,et al.  INCREASING TEST RELIABILITY THROUGH SELF-SCORING PROCEDURES , 1972 .

[15]  Kenneth D. Hopkins,et al.  Validity and Reliability Consequences of Confidence Weighting , 1973 .

[16]  R. Pugh,et al.  The Contribution of Selected Personality Traits and Knowledge to Response Behavior on a Probabilistic Test. , 1974 .

[17]  M. Kane,et al.  The Effect of SSM Grading on Reliability When Residual Items Have No Discriminating Power. , 1974 .

[18]  R. Koehler Overconfidence on Probabilistic Tests. , 1974 .

[19]  David J. Weiss,et al.  Final Report: Computerized Adaptive Ability Testing, April 1981 , 1981 .