Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2025 Apr 25:16:1522740.
doi: 10.3389/fpsyg.2025.1522740. eCollection 2025.

EuleApp©: a computerized adaptive assessment tool for early literacy skills

Affiliations

EuleApp©: a computerized adaptive assessment tool for early literacy skills

Melike Yumus et al. Front Psychol. .

Abstract

Introduction: Ample evidence indicates that assessing children's early literacy skills is crucial for later academic success. This assessment enables the provision of necessary support and materials while engaging them in the culture of print and books before school entry. However, relatively few assessment tools are available to identify early literacy skills, such as concepts of print, print awareness, phonological awareness, word awareness, alphabet knowledge, and early reading. The digital landscape presents new opportunities to enhance these assessments and provide enriching early literacy experiences. This study examines the psychometric properties of an adaptive assessment tool, EuLeApp©, focusing on its reliability and concurrent validity.

Methods: Data involved 307 German kindergarten children (Mage = 64 months old, range = 45-91). A Computerized Adaptive Testing (CAT) method, grounded in Item Response Theory (IRT), was employed to develop an adaptive digital tool for assessing early literacy competencies. We utilized an automatic item selection procedure based on item difficulty and discrimination parameters for the 183-item pool to ensure a precise and efficient assessment tailored to each child's ability level.

Results: The 4-parameter Logistic (4PL) model was identified as the best-fitting model for adaptive assessment, providing the highest precision in estimating children's abilities within this framework.

Discussions: The findings support the idea that the adaptive digital-based assessment tool EuLeApp© can be used to assess early literacy skills. It also provides a foundation for offering individualized and adaptable learning opportunities embedded in daily routines in daycare centers.

Keywords: computerized adaptive test; digital assessment; early literacy; item response theory; preschool age; psychometric validation.

PubMed Disclaimer

Conflict of interest statement

HL is the founder of DHL Data Science Seminars. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Figures

Figure 1
Figure 1
EuLeApp© subtests with sample test items.
Figure 2
Figure 2
The development of an adaptive app-based early literacy assessment tool.
Figure 3
Figure 3
A typical item characteristic curve for the 4PL IRT model. P(θ) represents the probability of a correct response given the ability level. χj (ability). This represents the ability level of the individual, typically ranging from very low to very high values. ϒj (probability of correct response). This represents the probability of answering the item correctly, ranging from 0 to 1. 4PL = four-parameter logistic.
Figure 4
Figure 4
Item difficulties in the calibrated item bank. MDISC: Item Discrimination Index, MDIFF: Item Difficulty Index. The MDIFF values were multiplied by −1 so that the difficulty of the items increases from left to right in the diagram.
Figure 5
Figure 5
Distribution of item selection per test based on stop rule. The figure shows the distribution of the number of items completed (x-axis labeled “number of items”) and their frequency (y-axis labeled “frequency”).
Figure 6
Figure 6
Test of t-scores across six scales using stop rule when standard errors (SE) < 0.0005. The solid lines represent the average t-scores, while the shaded areas indicate the standard error ranges.

References

    1. Ackerman T. A., Gierl M. J., Walker C. M. (2003). Using multidimensional item response theory to evaluate educational and psychological tests. Educ. Meas. Issues Pract. 22, 37–51. doi: 10.1111/j.1745-3992.2003.tb00136.x - DOI
    1. Adams M. J. (1994). Beginning to read: Thinking and learning about print.
    1. Antoniou F., Ralli A. M., Mouzaki A., Diamanti V., Papaioannou S. (2022). Logometro®: the psychometric properties of a norm-referenced digital battery for language assessment of Greek-speaking 4–7 years old children. Front. Psychol. 13:900600. doi: 10.3389/fpsyg.2022.900600, PMID: - DOI - PMC - PubMed
    1. Anderson V., Levin H. S., Jacobs R. (2002). Executive functions after frontal lobe injury: A developmental perspective. J. Clin. Exp. Neuropsychol. 24, 224–247. doi: 10.1207/S15327647JCD1201_02 - DOI
    1. Baker F. (2001). The basics of item response theory: ERIC clearinghouse on assessment and evaluation. College Park, MD: University of Maryland, ERIC Clearinghouse on Assessment and Evaluation.

LinkOut - more resources