Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2025 Apr 7;33(spe1):e279781.
doi: 10.1590/1413-785220243201e279781. eCollection 2025.

TROCHANTERIC FRACTURES: ARE THE CLASSIFICATIONS RELIABLE?

Affiliations

TROCHANTERIC FRACTURES: ARE THE CLASSIFICATIONS RELIABLE?

Luis Henrique Zambra Win et al. Acta Ortop Bras. .

Abstract

Objective: To evaluate classification use based on the assessment of agreement in trochanteric fractures between different observers through a simple proposal of bimodal classification.

Methodology: A total of 50 radiographic images of femur trochanteric fractures were selected and classified by 22 evaluators, 10 traumatologists and 12 residents, as STABLE or UNSTABLE. The assessment of reproducibility was done using the kappa statistical index. After the evaluation, the groups were isolated and submitted to the ANOVA test with Bonferroni correction to evaluate the statistical differences between them.

Results: When evaluated by the kappa index, the results of the 50 fractures assessed by the 22 evaluators were k = 0.272. The reproducibility of the classification proposal was considered statistically weak.

Conclusion: Based on this study, it is recommended that the classifications be extensively tested and reach a minimum level of interobserver reproducibility (kappa > 0.8). It is suggested that classifications that do not achieve this result be improved or abandoned. Level of evidence II, Prospective study.

Objetivo: Avaliar a utilização das classificações a partir da avaliação de concordância nas fraturas trocantéricas entre diferentes observadores através de uma proposta simples de classificação bimodal.

Métodos: Foram selecionadas 50 imagens radiográficas de fraturas trocantéricas do fêmur e classificadas por 22 avaliadores, sendo 10 traumatologistas e 12 residentes que definiram as fraturas como ESTÁVEIS ou INSTÁVEIS. A avaliação da reprodutibilidade foi dada pelo índice estatístico Kappa. Após a avaliação, os grupos foram isolados e submetidos ao teste ANOVA com correção Bonferroni para avaliar as diferenças estatísticas entre eles.

Resultados: Quando submetidos a avaliação pelo índice Kappa os resultados das 50 fraturas avaliadas pelos 22 avaliadores foi de k = 0,272. A reprodutibilidade da proposta de classificação foi considerada estatisticamente fraca.

Conclusão: A partir deste estudo recomenda-se que as classificações sejam extensamente testadas e atinjam um nível mínimo de reprodutibilidade (Kappa > 0,8) interobservadores. Sugere-se que as classificações que não atinjam este resultado sejam aprimoradas ou abandonadas. Nível de evidencia II, Estudo prospectivo.

Keywords: Classification; Interobserver agreement; Reliability; Trochanteric fracture.

PubMed Disclaimer

Conflict of interest statement

The authors declare no conflict of interest.

Figures

Figure 1
Figure 1. Image of the 50 radiographs used for the assessment.
Graph 1
Graph 1. Histogram of the variation and concentration of kappa indices among traumatologists
Graph 2
Graph 2. Histogram of the variation and concentration of kappa indices among R3
Graph 3
Graph 3. Histogram of the variation and concentration of kappa indices among R1 + R2
Graph 4
Graph 4. Histogram of the variation and concentration of kappa indices among all observers
Graph 5
Graph 5. Graph presenting each group’s kappa coefficient with 95% CI, demonstrating that the traumatologists group had a statistically significant difference in relation to the other groups (p = 0.007)

References

    1. Gwet KL. Large-Sample Variance of Fleiss Generalized Kappa. Educ Psychol Meas. 2021;81(4):781–790. - PMC - PubMed
    1. McHugh ML. Interrater reliability: the kappa statistic. Biochem Med. 2012;22(3):276–282. - PMC - PubMed
    1. Everitt BS. The analysis of contingency tables. 2. New York: Chapman & Hall; 1992.
    1. Cohen J. A coefficient of agreement for nominal scales. Educ Psychol Meas. 1960;20(1):37–46.
    1. Svanholm H, Starklint H, Gundersen HJ, Fabricius J, Barlebo H, Olsen S. Reproducibility of histomorphologic diagnoses with special reference to the kappa statistic. APMIS. 1989;97(8):689–698. - PubMed

LinkOut - more resources