Using Interpretable Machine Learning for Differential Item Functioning Detection in Psychometric Tests
- PMID: 39055539
- PMCID: PMC11268249
- DOI: 10.1177/01466216241238744
Using Interpretable Machine Learning for Differential Item Functioning Detection in Psychometric Tests
Abstract
This study presents a novel method to investigate test fairness and differential item functioning combining psychometrics and machine learning. Test unfairness manifests itself in systematic and demographically imbalanced influences of confounding constructs on residual variances in psychometric modeling. Our method aims to account for resulting complex relationships between response patterns and demographic attributes. Specifically, it measures the importance of individual test items, and latent ability scores in comparison to a random baseline variable when predicting demographic characteristics. We conducted a simulation study to examine the functionality of our method under various conditions such as linear and complex impact, unfairness and varying number of factors, unfair items, and varying test length. We found that our method detects unfair items as reliably as Mantel-Haenszel statistics or logistic regression analyses but generalizes to multidimensional scales in a straight forward manner. To apply the method, we used random forests to predict migration backgrounds from ability scores and single items of an elementary school reading comprehension test. One item was found to be unfair according to all proposed decision criteria. Further analysis of the item's content provided plausible explanations for this finding. Analysis code is available at: https://osf.io/s57rw/?view_only=47a3564028d64758982730c6d9c6c547.
Keywords: differential item functioning; interpretable machine learning; machine learning; psychometrics; random forest; test fairness.
© The Author(s) 2024.
Conflict of interest statement
The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.
Figures









References
-
- Attali Y., Bar‐Hillel M. (2003). Guess where: The position of correct answers in multiple‐choice test items as a psychometric variable. Journal of Educational Measurement, 40(2), 109–128. 10.1111/j.1745-3984.2003.tb01099.x - DOI
-
- Bauer D., Belzak W., Cole V. (2020). Simplifying the assessment of measurement invariance over multiple background variables: Using regularized moderated nonlinear factor analysis to detect differential item functioning. Structural Equation Modeling: A Multidisciplinary Journal, 27(1), 43–55. 10.1080/10705511.2019.1642754 - DOI - PMC - PubMed
-
- Becker B., Klein O., Biedinger N. (2013). The development of cognitive, language, and cultural skills from age 3 to 6: A comparison between children of Turkish origin and children of native-born German parents and the role of immigrant parents’ acculturation to the receiving society. American Educational Research Journal, 50(3), 616–649. 10.3102/0002831213480825 - DOI
-
- Belzak W. C. (2022). The multidimensionality of measurement bias in high‐stakes testing: Using machine learning to evaluate complex sources of differential item functioning. Educational Measurement: Issues and Practice, 42(1), 24–33. 10.1111/emip.12486 - DOI