Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2014 Jun 2;9(6):e96814.
doi: 10.1371/journal.pone.0096814. eCollection 2014.

Advancing bag-of-visual-words representations for lesion classification in retinal images

Affiliations

Advancing bag-of-visual-words representations for lesion classification in retinal images

Ramon Pires et al. PLoS One. .

Abstract

Diabetic Retinopathy (DR) is a complication of diabetes that can lead to blindness if not readily discovered. Automated screening algorithms have the potential to improve identification of patients who need further medical attention. However, the identification of lesions must be accurate to be useful for clinical application. The bag-of-visual-words (BoVW) algorithm employs a maximum-margin classifier in a flexible framework that is able to detect the most common DR-related lesions such as microaneurysms, cotton-wool spots and hard exudates. BoVW allows to bypass the need for pre- and post-processing of the retinographic images, as well as the need of specific ad hoc techniques for identification of each type of lesion. An extensive evaluation of the BoVW model, using three large retinograph datasets (DR1, DR2 and Messidor) with different resolution and collected by different healthcare personnel, was performed. The results demonstrate that the BoVW classification approach can identify different lesions within an image without having to utilize different algorithms for each lesion reducing processing time and providing a more flexible diagnostic system. Our BoVW scheme is based on sparse low-level feature detection with a Speeded-Up Robust Features (SURF) local descriptor, and mid-level features based on semi-soft coding with max pooling. The best BoVW representation for retinal image classification was an area under the receiver operating characteristic curve (AUC-ROC) of 97.8% (exudates) and 93.5% (red lesions), applying a cross-dataset validation protocol. To assess the accuracy for detecting cases that require referral within one year, the sparse extraction technique associated with semi-soft coding and max pooling obtained an AUC of 94.2 ± 2.0%, outperforming current methods. Those results indicate that, for retinal image classification tasks in clinical practice, BoVW is equal and, in some instances, surpasses results obtained using dense detection (widely believed to be the best choice in many vision problems) for the low-level descriptors.

PubMed Disclaimer

Conflict of interest statement

Competing Interests: The authors have the following interests: This work was supported in part by a research grant (2009–2011) from Microsoft Research (http://research.microsoft.com). It was an agreement between Microsoft Research and the São Paulo Research Foundation FAPESP in which they funded interesting research through the Microsoft Research–FAPESP Institute for IT Research (http://www.fapesp.br/en/5392). The institute supports high-quality fundamental research in information and communication technologies that is geared towards addressing social and economic development needs of the region. This work was also supported in part by Samsung Eletrônica da Amazônia (http://www.samsung.com). It is a scholarship program established by Samsung and our institution (Institute of Computing – http://ic.unicamp.br) to finance hard-working students. This does not alter the authors’ adherence to all the PLoS ONE policies on sharing data and materials.

Figures

Figure 1
Figure 1. The BoVW model illustrated as a matrix.
The figure highlights the relationship between the low-level features x j, the codewords c m of the visual dictionary, the encoded features αm, the coding function f and the pooling function g.
Figure 2
Figure 2. Regions of interest (dashed black regions) and the points of interest (blue circles).
Points of interest falling within the regions marked by the specialist are considered for creating the class-aware codebook – half of the codebook is learned from local features sampled inside the regions marked as lesions, and half the codebook is learned from local features outside those regions.
Figure 3
Figure 3. Standardized AUCs per lesion, for six combinations of feature extraction and coding (horizontal axis).
In the box-plots (black), the whiskers show the range up to 1.5× the interquartile range, and outliers are shown as small circles. Averages (small squares) and 95%-confidence intervals (error bars) are also shown, in red, for the same data. The strong synergy between sparse feature extraction and semi-soft coding is evident: it has consistently improved results for all lesions, while the other combinations improve the results of some lesions at the cost of decreasing it for other lesions (as shown by the spread of the standardized effects in the vertical axis). This plot is based on a balanced design with the DR2 dataset and all lesions, the other balanced design with both datasets and two lesions show similar results.
Figure 4
Figure 4. Final decision for necessity of referral.
The decision is based upon meta-classification using the scores of the individual lesion detectors as features. The meta-classifier is trained and tested on the DR2 dataset.

References

    1. Sinthanayothin C, Boyce JF, Williamson TH, Cook HL, Mensah E, et al. (2002) Automated detection of diabetic retinopathy on digital fundus images. Diabetic Medicine: a Journal of the British Diabetic Association 19: 105–112. - PubMed
    1. Jelinek HF, Cree MJ, Worsley D, Luckie AP, Nixon P (2006) An automated microaneurysm detector as a tool for identification of diabetic retinopathy in rural optometric practice. Clinical and Experimental Optometry 89: 299–305. - PubMed
    1. Fleming AD, Philip S, Goatman KA, Olson JA, Sharp PF (2006) Automated microaneurysm detection using local contrast normalization and local vessel detection. IEEE Transactions on Medical Imaging 25: 1223–1232. - PubMed
    1. Niemeijer M, van Ginneken B, Russell SR, Suttorp-Schulten MSA, Abràmoff MD (2007) Automated detection and differentiation of drusen, exudates, and cotton-wool spots in digital color fundus photographs for diabetic retinopathy diagnosis. Investigative Ophthalmology & Visual Science 48: 2260–2267. - PMC - PubMed
    1. Giancardo L, Mériaudeau F, Karnowski TP, Tobin KW, Li Y, et al... (2010) Microaneurysms Detection with the Radon Cliff Operator in Retinal Fundus Images. In SPIE Medical Imaging. 76230U-76230U. International Society for Optics and Photonics.

Publication types