Artificial Intelligence for Detecting Acute Fractures in Patients Admitted to an Emergency Department: Real-Life Performance of Three Commercial Algorithms
- PMID: 37468377
- DOI: 10.1016/j.acra.2023.06.016
Artificial Intelligence for Detecting Acute Fractures in Patients Admitted to an Emergency Department: Real-Life Performance of Three Commercial Algorithms
Abstract
Rationale and objectives: Interpreting radiographs in emergency settings is stressful and a burden for radiologists. The main objective was to assess the performance of three commercially available artificial intelligence (AI) algorithms for detecting acute peripheral fractures on radiographs in daily emergency practice.
Materials and methods: Radiographs were collected from consecutive patients admitted for skeletal trauma at our emergency department over a period of 2 months. Three AI algorithms-SmartUrgence, Rayvolve, and BoneView-were used to analyze 13 body regions. Four musculoskeletal radiologists determined the ground truth from radiographs. The diagnostic performance of the three AI algorithms was calculated at the level of the radiography set. Accuracies, sensitivities, and specificities for each algorithm and two-by-two comparisons between algorithms were obtained. Analyses were performed for the whole population and for subgroups of interest (sex, age, body region).
Results: A total of 1210 patients were included (mean age 41.3 ± 18.5 years; 742 [61.3%] men), corresponding to 1500 radiography sets. The fracture prevalence among the radiography sets was 23.7% (356/1500). Accuracy was 90.1%, 71.0%, and 88.8% for SmartUrgence, Rayvolve, and BoneView, respectively; sensitivity 90.2%, 92.6%, and 91.3%, with specificity 92.5%, 70.4%, and 90.5%. Accuracy and specificity were significantly higher for SmartUrgence and BoneView than Rayvolve for the whole population (P < .0001) and for subgroups. The three algorithms did not differ in sensitivity (P = .27). For SmartUrgence, subgroups did not significantly differ in accuracy, specificity, or sensitivity. For Rayvolve, accuracy and specificity were significantly higher with age 27-36 than ≥53 years (P = .0029 and P = .0019). Specificity was higher for the subgroup knee than foot (P = .0149). For BoneView, accuracy was significantly higher for the subgroups knee than foot (P = .0006) and knee than wrist/hand (P = .0228). Specificity was significantly higher for the subgroups knee than foot (P = .0003) and ankle than foot (P = .0195).
Conclusion: The performance of AI detection of acute peripheral fractures in daily radiological practice in an emergency department was good to high and was related to the AI algorithm, patient age, and body region examined.
Keywords: Artificial intelligence; Deep learning; Fracture; Medical imaging; Musculoskeletal imaging.
Copyright © 2023 The Association of University Radiologists. Published by Elsevier Inc. All rights reserved.
Conflict of interest statement
Declaration of Competing Interest Valérie Bousson is a paid consultant for Milvue starting July 1, 2022. Grégoire Attané was paid by Gleamer between March and August 2019 to label radiographs. The remaining authors declare none. The three companies, Milvue, AZmed, and Gleamer, provided the AI algorithms to our radiology department for free. None of the three companies had access to the study methodology or data during the course of the study.
Similar articles
-
Commercially-available AI algorithm improves radiologists' sensitivity for wrist and hand fracture detection on X-ray, compared to a CT-based ground truth.Eur Radiol. 2024 May;34(5):2885-2894. doi: 10.1007/s00330-023-10380-1. Epub 2023 Nov 3. Eur Radiol. 2024. PMID: 37919408
-
Comparative accuracy of two commercial AI algorithms for musculoskeletal trauma detection in emergency radiographs.Emerg Radiol. 2025 Aug;32(4):569-580. doi: 10.1007/s10140-025-02353-2. Epub 2025 Jun 9. Emerg Radiol. 2025. PMID: 40489014 Free PMC article.
-
Comparison of diagnostic performance of a deep learning algorithm, emergency physicians, junior radiologists and senior radiologists in the detection of appendicular fractures in children.Pediatr Radiol. 2023 Jul;53(8):1675-1684. doi: 10.1007/s00247-023-05621-w. Epub 2023 Mar 6. Pediatr Radiol. 2023. PMID: 36877239
-
Artificial Intelligence (AI) for Fracture Diagnosis: An Overview of Current Products and Considerations for Clinical Adoption, From the AJR Special Series on AI Applications.AJR Am J Roentgenol. 2022 Dec;219(6):869-878. doi: 10.2214/AJR.22.27873. Epub 2022 Jun 22. AJR Am J Roentgenol. 2022. PMID: 35731103 Review.
-
Artificial intelligence diagnostic accuracy in fracture detection from plain radiographs and comparing it with clinicians: a systematic review and meta-analysis.Clin Radiol. 2024 Aug;79(8):579-588. doi: 10.1016/j.crad.2024.04.009. Epub 2024 Apr 20. Clin Radiol. 2024. PMID: 38772766
Cited by
-
Diagnostic performance of ChatGPT in tibial plateau fracture in knee X-ray.Emerg Radiol. 2025 Feb;32(1):59-64. doi: 10.1007/s10140-024-02298-y. Epub 2024 Nov 30. Emerg Radiol. 2025. PMID: 39613920
-
Impact of deep learning on pediatric elbow fracture detection: a systematic review and meta-analysis.Eur J Trauma Emerg Surg. 2025 Feb 20;51(1):115. doi: 10.1007/s00068-025-02779-w. Eur J Trauma Emerg Surg. 2025. PMID: 39976732
-
Diagnostic value of artificial intelligence-based software for the detection of pediatric upper extremity fractures.Eur Radiol. 2025 Aug 23. doi: 10.1007/s00330-025-11947-w. Online ahead of print. Eur Radiol. 2025. PMID: 40848140
-
Optimizing the power of AI for fracture detection: from blind spots to breakthroughs.Skeletal Radiol. 2025 Oct;54(10):2007-2024. doi: 10.1007/s00256-025-04951-0. Epub 2025 May 23. Skeletal Radiol. 2025. PMID: 40407826 Review.
-
Cross-validation of an artificial intelligence tool for fracture classification and localization on conventional radiography in Dutch population.Insights Imaging. 2025 Jul 3;16(1):150. doi: 10.1186/s13244-025-02034-1. Insights Imaging. 2025. PMID: 40610833 Free PMC article.
MeSH terms
LinkOut - more resources
Full Text Sources
Medical