Assessment Scores of a Mock Objective Structured Clinical Examination Administered to 99 Anesthesiology Residents at 8 Institutions
- PMID: 32149757
- DOI: 10.1213/ANE.0000000000004705
Assessment Scores of a Mock Objective Structured Clinical Examination Administered to 99 Anesthesiology Residents at 8 Institutions
Abstract
Background: Objective Structured Clinical Examinations (OSCEs) are used in a variety of high-stakes examinations. The primary goal of this study was to examine factors influencing the variability of assessment scores for mock OSCEs administered to senior anesthesiology residents.
Methods: Using the American Board of Anesthesiology (ABA) OSCE Content Outline as a blueprint, scenarios were developed for 4 of the ABA skill types: (1) informed consent, (2) treatment options, (3) interpretation of echocardiograms, and (4) application of ultrasonography. Eight residency programs administered these 4 OSCEs to CA3 residents during a 1-day formative session. A global score and checklist items were used for scoring by faculty raters. We used a statistical framework called generalizability theory, or G-theory, to estimate the sources of variation (or facets), and to estimate the reliability (ie, reproducibility) of the OSCE performance scores. Reliability provides a metric on the consistency or reproducibility of learner performance as measured through the assessment.
Results: Of the 115 total eligible senior residents, 99 participated in the OSCE because the other residents were unavailable. Overall, residents correctly performed 84% (standard deviation [SD] 16%, range 38%-100%) of the 36 total checklist items for the 4 OSCEs. On global scoring, the pass rate for the informed consent station was 71%, for treatment options was 97%, for interpretation of echocardiograms was 66%, and for application of ultrasound was 72%. The estimate of reliability expressing the reproducibility of examinee rankings equaled 0.56 (95% confidence interval [CI], 0.49-0.63), which is reasonable for normative assessments that aim to compare a resident's performance relative to other residents because over half of the observed variation in total scores is due to variation in examinee ability. Phi coefficient reliability of 0.42 (95% CI, 0.35-0.50) indicates that criterion-based judgments (eg, pass-fail status) cannot be made. Phi expresses the absolute consistency of a score and reflects how closely the assessment is likely to reproduce an examinee's final score. Overall, the greatest (14.6%) variance was due to the person by item by station interaction (3-way interaction) indicating that specific residents did well on some items but poorly on other items. The variance (11.2%) due to residency programs across case items was high suggesting moderate variability in performance from residents during the OSCEs among residency programs.
Conclusions: Since many residency programs aim to develop their own mock OSCEs, this study provides evidence that it is possible for programs to create a meaningful mock OSCE experience that is statistically reliable for separating resident performance.
Similar articles
-
A Descriptive Survey of Anesthesiology Residency Simulation Programs: How Are Programs Preparing Residents for the New American Board of Anesthesiology APPLIED Certification Examination?Anesth Analg. 2017 Sep;125(3):991-998. doi: 10.1213/ANE.0000000000002189. Anesth Analg. 2017. PMID: 28632531
-
Effect of clinically discriminating, evidence-based checklist items on the reliability of scores from an Internal Medicine residency OSCE.Adv Health Sci Educ Theory Pract. 2014 Oct;19(4):497-506. doi: 10.1007/s10459-013-9482-4. Epub 2014 Jan 22. Adv Health Sci Educ Theory Pract. 2014. PMID: 24449122
-
Development of an Objective Structured Clinical Examination Using the American Board of Anesthesiology Content Outline for the Objective Structured Clinical Examination Component of the APPLIED Certification Examination.A A Pract. 2018 Oct 1;11(7):193-197. doi: 10.1213/XAA.0000000000000779. A A Pract. 2018. PMID: 29688921
-
Potential Harm of Elimination of Score Reporting for the United States Medical Licensing Examination Step 1 Examination to Anesthesiology Residency Selection.Anesth Analg. 2021 Jan;132(1):275-279. doi: 10.1213/ANE.0000000000005235. Anesth Analg. 2021. PMID: 33031347 Review. No abstract available.
-
A systematic review and meta-analysis of measurement properties of objective structured clinical examinations used in physical therapy licensure and a structured review of licensure practices in countries with well-developed regulation systems.PLoS One. 2021 Aug 3;16(8):e0255696. doi: 10.1371/journal.pone.0255696. eCollection 2021. PLoS One. 2021. PMID: 34343213 Free PMC article.
Cited by
-
Reliability of simulation-based assessment for practicing physicians: performance is context-specific.BMC Med Educ. 2021 Apr 12;21(1):207. doi: 10.1186/s12909-021-02617-8. BMC Med Educ. 2021. PMID: 33845837 Free PMC article.
-
Simulation-Based Education in Acute and Chronic Pain Training.Curr Pain Headache Rep. 2023 Nov;27(11):639-643. doi: 10.1007/s11916-023-01164-9. Epub 2023 Sep 16. Curr Pain Headache Rep. 2023. PMID: 37715889 Review.
References
-
- Kogan JR, Holmboe ES, Hauer KETools for direct observation and assessment of clinical skills of medical trainees: a systematic review. JAMA. 2009;302:1316–1326.
-
- Hastie MJ, Spellman JL, Pagano PP, Hastie J, Egan BJDesigning and implementing the objective structured clinical examination in anesthesiology. Anesthesiology. 2014;120:196–203.
-
- Harden RM, Stevenson M, Downie WW, Wilson GMAssessment of clinical competence using objective structured examination. Br Med J. 1975;1:447–451.
-
- Yamamoto S, Tanaka P, Madsen MV, Macario AComparing anesthesiology residency training structure and requirements in seven different countries on three continents. Cureus. 2017;9:e1060.
-
- Bromley LMThe objective structured clinical exam - practical aspects. Curr Opin Anaesthesiol. 2000;13:675–678.
Publication types
MeSH terms
LinkOut - more resources
Full Text Sources
Miscellaneous