Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
Review
. 2004 Sep;19(9):971-7.
doi: 10.1111/j.1525-1497.2004.40066.x.

How reliable are assessments of clinical teaching? A review of the published instruments

Affiliations
Review

How reliable are assessments of clinical teaching? A review of the published instruments

Thomas J Beckman et al. J Gen Intern Med. 2004 Sep.

Abstract

Background: Learner feedback is the primary method for evaluating clinical faculty, despite few existing standards for measuring learner assessments.

Objective: To review the published literature on instruments for evaluating clinical teachers and to summarize themes that will aid in developing universally appealing tools.

Design: Searching 5 electronic databases revealed over 330 articles. Excluded were reviews, editorials, and qualitative studies. Twenty-one articles describing instruments designed for evaluating clinical faculty by learners were found. Three investigators studied these papers and tabulated characteristics of the learning environments and validation methods. Salient themes among the evaluation studies were determined.

Main results: Many studies combined evaluations from both outpatient and inpatient settings and some authors combined evaluations from different learner levels. Wide ranges in numbers of teachers, evaluators, evaluations, and scale items were observed. The most frequently encountered statistical methods were factor analysis and determining internal consistency reliability with Cronbach's alpha. Less common methods were the use of test-retest reliability, interrater reliability, and convergent validity between validated instruments. Fourteen domains of teaching were identified and the most frequently studied domains were interpersonal and clinical-teaching skills.

Conclusions: Characteristics of teacher evaluations vary between educational settings and between different learner levels, indicating that future studies should utilize more narrowly defined study populations. A variety of validation methods including temporal stability, interrater reliability, and convergent validity should be considered. Finally, existing data support the validation of instruments comprised solely of interpersonal and clinical-teaching domains.

PubMed Disclaimer

Figures

FIGURE 1
FIGURE 1
Domains of teaching: frequency of use in evaluation instruments (N = 21 instruments).

References

    1. Jones RG, Froom JD. Faculty and administration view of problems in faculty evaluations. Acad Med. 1994;69:476–83. - PubMed
    1. Beckman TJ, Lee MC, Mandrekar JN. A comparison of clinical teaching evaluations by resident and peer physicians. Med Teach. 2004;26:321–5. - PubMed
    1. Downing SM. Validity: on the meaningful interpretation of assessment data. Med Educ. 2003;37:830–7. - PubMed
    1. Crossley J, Humphris G, Jolly B. Assessing health professionals. Med Educ. 2002;36:800–4. - PubMed
    1. Beckman TJ, Lee MC, Rohren CH, Pankratz VS. Evaluating an instrument for the peer review of inpatient teaching. Med Teach. 2003;25:131–5. - PubMed