Abstract

The validity of a standardised test is questioned if an irrelevant construct is accounted for the performance of examinees, which is wrongly modeled as the ability in the construct (test items). A test must ensure precision in the examinee's ability irrespective of their sub-population in any demographic variables. This paper explored the potentials of gender and school location as major covariates on the West African Examinations Council (WAEC) mathematics items among examinees (N=2,866) using Exploratory Structural Equation Modeling (ESEM). The results remarked that the test is multidimensional (six-factors) with compliance fix indices of (χ2 (940)=4882.024, p < 0.05, CFI=0.962, TLI=0.930, RMSEA=0.038, SRMR=0.030, 90 % CI=0.037-0.039, Akaike information criterion (AIC)=147290.577, Bayesian information criterion (BIC)=149585.436 and Sample-size adjusted BIC=148362.154) respectively. Also, there were 10 (20 %) significant DIF items in the WAEC to gender, while 3 (6 %) of the items indicated significant DIF to school location. Observed DIF items acquaint test developers; the existence of DIF may differentially affect the performance of examinees with the same ability level. The implications of the test are severe for the examinees. Hence, accurate and unbiased assessment should be the basic principles for any test item measurement, and test developers need to test the items to be free from biases psychometrically.

Highlights

  • Ensuring test fairness and equity among examinees is very important

  • When the fitness of 2-factor and 3-factor models was compared, the analysis showed that 3-factor fitted the data better than the 2-factor model with χ2 (48)=920.43, p

  • It’s crucial for item commission writers and test developers of this public examining body to ensure that the test items are valid, reliable, and free from bias

Read more

Summary

Introduction

Ensuring test fairness and equity among examinees is very important. The measurement community has increasingly employed standardised testing as one of the significant tools, deployed in assessing examinees’ outcomes and precise ability and examinee potential for future academic success. Measurement experts over the years were apprehensive of the likelihood that test items, either cognitive or non-cognitive, might function differently (that is, favouring sub-group) across a group of examinees [1, 2]. The validity of a standardised test is questioned if an irrelevant construct is accounted for the performance of examinees, which is wrongly modeled as the ability in the construct (test items). A test must ensure precision in the examinee’s ability irrespective of their sub-population in any demographic variables. The examinee’s test score interpretation concerns the extent of its statistical independence across and among different groups of examinees in educational testing

Objectives
Methods
Results
Conclusion
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.