Abstract
Scoring constructed response items is more time consuming and complex than scoring multiple-choice items. Many educational and noneducational institutions adapt open-ended questions in examinations for admission, certification, graduation, accountability, and licensing purposes. These examinations often are administered on a large-scale basis. Large volumes of examinees are tested and yet only a short time (usually a few weeks) is available for scoring. Many raters are recruited to score the examinations and it is infeasible to assign all the raters to score every one of the examinees. Thus each examinee will be scored by a selection of raters, leading to sparsely-filled data sets and also unbalanced designs, and also causing potentially biased and imprecise estimators. The current monograph developed and validated a method, called the subdividing method, to resolve this problem. The subdividing method, drawing on the concept that one could obtain more stable estimators by synthesizing multiple data sources than using just one source, is set out to improve the accuracy and precision of estimates quantifying measurement errors in the framework of G theory. The implementation of this method was discussed in section 3.1.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 2001 Springer Science+Business Media New York
About this chapter
Cite this chapter
Chiu, C.WT. (2001). Conclusions, Discussions, and Future Directions. In: Scoring Performance Assessments Based on Judgements. Evaluation in Education and Human Services, vol 50. Springer, Dordrecht. https://doi.org/10.1007/978-94-010-0650-7_5
Download citation
DOI: https://doi.org/10.1007/978-94-010-0650-7_5
Publisher Name: Springer, Dordrecht
Print ISBN: 978-94-010-3871-3
Online ISBN: 978-94-010-0650-7
eBook Packages: Springer Book Archive