Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2015 Jul;42(7):4241-9.
doi: 10.1118/1.4922681.

Assessment of performance and reproducibility of applying a content-based image retrieval scheme for classification of breast lesions

Affiliations

Assessment of performance and reproducibility of applying a content-based image retrieval scheme for classification of breast lesions

Rohith Reddy Gundreddy et al. Med Phys. 2015 Jul.

Abstract

Purpose: To develop a new computer-aided diagnosis (CAD) scheme using a content-based image retrieval (CBIR) approach for classification between the malignant and benign breast lesions depicted on the digital mammograms and assess CAD performance and reproducibility.

Methods: An image dataset including 820 regions of interest (ROIs) was used. Among them, 431 ROIs depict malignant lesions and 389 depict benign lesions. After applying an image preprocessing process to define the lesion center, two image features were computed from each ROI. The first feature is an average pixel value of a mapped region generated using a watershed algorithm. The second feature is an average pixel value difference between a ROI's center region and the rest of the image. A two-step CBIR approach uses these two features sequentially to search for ten most similar reference ROIs for each queried ROI. A similarity based classification score was then computed to predict the likelihood of the queried ROI depicting a malignant lesion. To assess the reproducibility of the CAD scheme, we selected another independent testing dataset of 100 ROIs. For each ROI in the testing dataset, we added four randomly queried lesion center pixels and examined the variation of the classification scores.

Results: The area under the ROC curve (AUC) = 0.962 ± 0.006 was obtained when applying a leave-one-out validation method to 820 ROIs. Using the independent testing dataset, the initial AUC value was 0.832 ± 0.040, and using the median classification score of each ROI with five queried seeds, AUC value increased to 0.878 ± 0.035.

Conclusions: The authors demonstrated that (1) a simple and efficient CBIR scheme using two lesion density distribution related features achieved high performance in classifying breast lesions without actual lesion segmentation and (2) similar to the conventional CAD schemes using global optimization approaches, improving reproducibility is also one of the challenges in developing CAD schemes using a CBIR based regional optimization approach.

PubMed Disclaimer

Figures

FIG. 1.
FIG. 1.
Illustration of 50 malignant ROIs extracted in our independent testing dataset.
FIG. 2.
FIG. 2.
Illustration of 50 benign ROIs extracted in our independent testing dataset.
FIG. 3.
FIG. 3.
An example of a matrix extracted from “a lesion center” (a) and the corresponding output of applying watershed function processing (b).
FIG. 4.
FIG. 4.
A diagram showing the variation of AUC values versus the increase of kernel size of a watershed matrix to compute feature 1 (F1).
FIG. 5.
FIG. 5.
A scatter diagram showing the distribution of two image features computed from 820 ROIs. The red square marks indicate malignant ROIs, and the blue diamond marks represent benign ROIs.
FIG. 6.
FIG. 6.
Histograms of the number of malignant reference ROIs retrieved using feature 1 (F1) among the 50 malignant and 50 benign testing ROIs.
FIG. 7.
FIG. 7.
Comparison of three ROC curves generated using the rockit program. The computed AUC values are 0.962 ± 0.006, 0.603 ± 0.020, and 0.515 ± 0.020 when using our CBIR based classification scores (a leave-one-ROI-out based validation method), the average of two image features, and the first feature computed from the watershed algorithm generated maps, respectively.
FIG. 8.
FIG. 8.
Distribution of classification scores computed using the original queried lesion center seeds (marked by *) and the median classification scores (marked by ) among 50 malignant ROIs (a) and 50 benign ROIs (b). In both diagrams (a) and (b), the solid red line and dashed blue line indicate the average classification score level of the 50 ROIs using the original seeds and median classification scores using the five seeds, respectively.
FIG. 9.
FIG. 9.
Comparison of two ROC curves generated using two different sets of classification scores of 100 testing ROIs. In these two ROC curves, AUC = 0.832 ± 0.040 and 0.878 ± 0.035 when using the classification scores computed based on the original queried lesion center seed and the median classification scores computed from five randomly placed lesion center seeds, respectively.

Similar articles

Cited by

References

    1. Siegel R., Naishadham D., and Jemal A., “Cancer statistics, 2013,” Ca-Cancer J. Clin. 63, 11–30 (2013).10.3322/caac.21166 - DOI - PubMed
    1. Cady B. and Michaelson J. S., “The life-sparing potential of mammographic screening,” Cancer 91, 1699–1703 (2001).10.1002/1097-0142(20010501)91:9<1699::AID-CNCR1186>3.0.CO;2-W - DOI - PubMed
    1. American College of Radiology , ACR Breast Imaging Reporting and Data System Atlas (BI-RADS Atlas) (American College of Radiology, Reston, VA, 2003).
    1. Berg W. A. et al., “Diagnostic accuracy of mammography, clinical examination, US, and MR imaging in preoperative assessment of breast cancer,” Radiology 233, 830–849 (2004).10.1148/radiol.2333031484 - DOI - PubMed
    1. Sickles E. A., Wolverton D. E., and Dee K. E., “Performance parameters for screening and diagnostic mammography: Specialist and general radiologists,” Radiology 224, 861–869 (2002).10.1148/radiol.2243011482 - DOI - PubMed

Publication types