Understanding Scoring Reliability: Experiments in Calibrating Essay Readers
Henry I. Braun
Journal of Educational and Behavioral Statistics, 1988, vol. 13, issue 1, 1-18
Abstract:
Scoring reliability of essays and other free-response questions is of considerable concern, especially in large, national administrations. This report describes a statistically designed experiment that was carried out in an operational setting to determine the contributions of different sources of variation to the unreliability of scoring. The experiment made novel use of partially balanced incomplete block designs that facilitated the unbiased estimation of certain main effects without requiring readers to assess the same paper several times. In addition, estimates were obtained of the improvement in reliability that results from removing variability from systematic sources of variation by an appropriate adjustment of the raw scores. This statistical calibration appears to be a cost-effective approach to enhancing scoring reliability when compared to simply increasing the number of readings per paper. The results of the experiment also provide a framework for examining other, simpler calibration strategies. One such strategy is briefly considered.
Keywords: reliability; designed experiments; essay scoring (search for similar items in EconPapers)
Date: 1988
References: Add references at CitEc
Citations:
Downloads: (external link)
https://journals.sagepub.com/doi/10.3102/10769986013001001 (text/html)
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:sae:jedbes:v:13:y:1988:i:1:p:1-18
DOI: 10.3102/10769986013001001
Access Statistics for this article
More articles in Journal of Educational and Behavioral Statistics
Bibliographic data for series maintained by SAGE Publications ().