Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2022 Feb;75(1):136-157.
doi: 10.1111/bmsp.12252. Epub 2021 Aug 30.

Computerized adaptive testing for testlet-based innovative items

Affiliations

Computerized adaptive testing for testlet-based innovative items

Hyeon-Ah Kang et al. Br J Math Stat Psychol. 2022 Feb.

Abstract

Increasing use of innovative items in operational assessments has shedded new light on the polytomous testlet models. In this study, we examine performance of several scoring models when polytomous items exhibit random testlet effects. Four models are considered for investigation: the partial credit model (PCM), testlet-as-a-polytomous-item model (TPIM), random-effect testlet model (RTM), and fixed-effect testlet model (FTM). The performance of the models was evaluated in two adaptive testings where testlets have nonzero random effects. The outcomes of the study suggest that, despite the manifest random testlet effects, PCM, FTM, and RTM perform comparably in trait recovery and examinee classification. The overall accuracy of PCM and FTM in trait inference was comparable to that of RTM. TPIM consistently underestimated population variance and led to significant overestimation of measurement precision, showing limited utility for operational use. The results of the study provide practical implications for using the polytomous testlet scoring models.

Keywords: adaptive testing; polytomous items; technology-enhanced innovative items; testlet.

PubMed Disclaimer

References

    1. Andrich, D. (1978). A rating formulation for ordered response categories. Psychometrika, 43, 561-573. https://doi.org/10.1007/BF02293814
    1. Andrich, D., & Marais, I. (2019). A course in Rasch measurement theory. Singapore, Singapore: Springer.
    1. Betts, J., Muntean, W., Kim, D., & Kao, S. (2021). Evaluating different scoring methods for multiple response items providing partial credit. Educational Psychological Measurement, 1-26. https://doi.org/10.1177/0013164421994636
    1. Bond, T., Yan, Z., & Heene, M. (2020). Applying the Rasch model: Fundamental measurement in the human sciences. New York, NY: Routledge. https://doi.org/10.4324/9780429030499
    1. Boyd, A. M., Dodd, B., & Fitzpatrick, S. (2013). A comparison of exposure control procedures in cat systems based on different measurement models for testlets. Applied Measurement in Education, 26, 113-135. https://doi.org/10.1080/08957347.2013.765434

MeSH terms

LinkOut - more resources