Competency-Based Assessments: Leveraging Artificial Intelligence to Predict Subcompetency Content
- PMID: 36477379
- DOI: 10.1097/ACM.0000000000005115
Competency-Based Assessments: Leveraging Artificial Intelligence to Predict Subcompetency Content
Abstract
Purpose: Faculty feedback on trainees is critical to guiding trainee progress in a competency-based medical education framework. The authors aimed to develop and evaluate a Natural Language Processing (NLP) algorithm that automatically categorizes narrative feedback into corresponding Accreditation Council for Graduate Medical Education Milestone 2.0 subcompetencies.
Method: Ten academic anesthesiologists analyzed 5,935 narrative evaluations on anesthesiology trainees at 4 graduate medical education (GME) programs between July 1, 2019, and June 30, 2021. Each sentence (n = 25,714) was labeled with the Milestone 2.0 subcompetency that best captured its content or was labeled as demographic or not useful. Inter-rater agreement was assessed by Fleiss' Kappa. The authors trained an NLP model to predict feedback subcompetencies using data from 3 sites and evaluated its performance at a fourth site. Performance metrics included area under the receiver operating characteristic curve (AUC), positive predictive value, sensitivity, F1, and calibration curves. The model was implemented at 1 site in a self-assessment exercise.
Results: Fleiss' Kappa for subcompetency agreement was moderate (0.44). Model performance was good for professionalism, interpersonal and communication skills, and practice-based learning and improvement (AUC 0.79, 0.79, and 0.75, respectively). Subcompetencies within medical knowledge and patient care ranged from fair to excellent (AUC 0.66-0.84 and 0.63-0.88, respectively). Performance for systems-based practice was poor (AUC 0.59). Performances for demographic and not useful categories were excellent (AUC 0.87 for both). In approximately 1 minute, the model interpreted several hundred evaluations and produced individual trainee reports with organized feedback to guide a self-assessment exercise. The model was built into a web-based application.
Conclusions: The authors developed an NLP model that recognized the feedback language of anesthesiologists across multiple GME programs. The model was operationalized in a self-assessment exercise. It is a powerful tool which rapidly organizes large amounts of narrative feedback.
References
-
- Frank JR, Mungroo R, Ahmad Y, Wang M, De Rossi S, Horsley T. Toward a definition of competency-based education in medicine: A systematic review of published definitions. Med Teach. 2010;32:631–637.
-
- Warm EJ, Edgar L, Kelleher M, et al. A Guidebook for Implementing and Changing Assessment in the Milestones Era. Accreditation Council for Graduate Medical Education. https://www.acgme.org/Portals/0/Milestones%20Implementation%202020.pdf . Published 2020. Accessed November 14, 2022.
-
- Powell DE, Carraccio C. Toward competency-based medical education. N Engl J Med. 2018;378:3–5.
-
- Kogan JR, Hatala R, Hauer KE, Guidelines HE. The do’s, don’ts and don’t knows of direct observation of clinical skills in medical education. Perspect Med Educ. 2017;6:286–305.
-
- Edgar L, McLean S, Hogan SO, Hamstra S, Holmboe ES. Accreditation Council for Graduate Medical Education (ACGME): The Milestones Guidebook. Chicago, IL: ACGME; 2020.
MeSH terms
LinkOut - more resources
Full Text Sources