Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2007 Mar;35(3):437-41.
doi: 10.1177/0363546506298108. Epub 2007 Jan 31.

Interobserver agreement in the classification of rotator cuff tears

Affiliations

Interobserver agreement in the classification of rotator cuff tears

John E Kuhn et al. Am J Sports Med. 2007 Mar.

Abstract

Background: Six classification systems have been proposed for describing rotator cuff tears designed to help understand their natural history and make treatment decisions.

Purpose: To assess the interobserver variation for these classification systems and identify the method with the best interobserver agreement.

Study design: Cohort study (diagnosis); Level of evidence, 2.

Methods: Six rotator cuff tear classification systems were identified in a literature search. The components of these systems included partial-thickness rotator cuff tears and classification by size, shape, configuration, number of tendons involved, and by extent, topography, and nature of the biceps. Twelve fellowship-trained orthopaedic surgeons who each perform at least 30 rotator cuff repairs per year reviewed arthroscopy videos from 30 patients with a random assortment of rotator cuff tears and classified them by the 6 classification systems. Interobserver variation was determined by a kappa analysis.

Results: Interobserver agreement was high when distinguishing between full-thickness and partial-thickness tears (0.95, kappa = 0.85). The investigators agreed on the side (articular vs bursal) of involvement for partial-thickness tears (observed agreement 0.92, kappa = 0.85) but could not agree when classifying the depth of the partial-thickness tear (observed agreement 0.49, kappa = 0.19). The best agreement for full-thickness tears was seen when the tear was classified by topography (degree of retraction) in the frontal plane (observed agreement 0.70, kappa = 0.54).

Conclusion: With the exception of distinguishing partial-thickness from full-thickness rotator cuff tears and identifying the side (articular vs bursal) of involvement with partial-thickness tears, currently described rotator cuff classification systems have little interobserver agreement among experienced shoulder surgeons. Researchers should consider describing full-thickness rotator cuff tears by topography (degree of retraction) in the frontal plane.

PubMed Disclaimer

Similar articles

Cited by

LinkOut - more resources