Abstract

Judge-mediated examinations are defined as those for which expert evaluation (using rubrics) is required to determine correctness, completeness, and reasonability of test-taker responses. The use of multifaceted Rasch modeling has led to improvements in the reliability of scoring such examinations. The establishment of criterion-referenced standards for these assessments has, however, remained a labor-intensive practice. This article explores a method for incorporating the establishment of minimal competency within the measurement model itself. Two high-stakes testing boards participated in a controlled experiment comparing a multifaceted Rasch modeled approach to a traditional modified borderline approach often used for standard setting in judge-mediated situations. Criterion points from each were compared to assess differences and support validity. Results of the experiment suggest that modeling standard setting as part of the performance evaluation is an effective and efficient way to define minimal competence.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call