Abstract

PurposeTo increase the precision of estimated item parameters of item response theory models for patient-reported outcomes, general population samples are often enriched with samples of clinical respondents. Calibration studies provide little information on how this sampling scheme is incorporated into model estimation. In a small simulation study the impact of ignoring the oversampling of clinical respondents on item and person parameters is illustrated.MethodSimulations were performed using two scenarios. Under the first it was assumed that regular and clinical respondents form two distinct distributions; under the second it was assumed that they form a single distribution. A synthetic item bank with quasi-trait characteristics was created, and item scores were generated from this bank for samples with varying percentages of clinical respondents. Proper (using a multi-group model, and sample weights, respectively, for Scenarios 1 and 2) and improper (ignoring oversampling) approaches for dealing with the clinical sample were contrasted using correlations and differences between true and estimated parameters.ResultsUnder the first scenario, ignoring the sampling scheme resulted in overestimation of both item and person parameters with bias decreasing with higher percentages of clinical respondents. Under the second, location and person parameters were underestimated with bias increasing in size with increasing percentage of clinical respondents. Under both scenarios, the standard error of the latent trait estimate was generally underestimated.ConclusionIgnoring the addition of extra clinical respondents leads to bias in item and person parameters, which may lead to biased norms and unreliable CAT scores. An appeal is made for researchers to provide more information on how clinical samples are incorporated in model estimation.

Highlights

  • In the last decade computerized adaptive testing (CAT) has become a popular method for efficiently assessing patientreported outcomes (PROs)

  • Ignoring the addition of extra clinical respondents leads to bias in item and person parameters, which may lead to biased norms and unreliable CAT scores

  • An appeal is made for researchers to provide more information on how clinical samples are incorporated in model estimation

Read more

Summary

Introduction

In the last decade computerized adaptive testing (CAT) has become a popular method for efficiently assessing patientreported outcomes (PROs). In the USA, the patient-reported outcomes measurement information system (PROMIS), a group of scientist from several academic institutions and the National Institutes of Health [1], has developed a multitude of CATs for the measurement of health status for physical, mental, and social well-being for use in clinical research and healthcare delivery settings [2]. Qual Life Res (2016) 25:1635–1644 representing item properties and patient characteristics. The patient characteristic is commonly referred to as ‘‘latent trait,’’ and denoted by parameter h. The latent trait scale is arbitrary by definition [8], and a linear transformation of a chosen scale results in identical expected item category probabilities. IRT software anchors the scale by putting it on a z-score scale, setting the mean and standard deviation of the latent trait to zero and one, respectively [9, Chap. IRT software anchors the scale by putting it on a z-score scale, setting the mean and standard deviation of the latent trait to zero and one, respectively [9, Chap. 6]

Objectives
Methods
Results
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call