Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2021 Dec 28:2021:2751695.
doi: 10.1155/2021/2751695. eCollection 2021.

Data Homogeneity Effect in Deep Learning-Based Prediction of Type 1 Diabetic Retinopathy

Affiliations

Data Homogeneity Effect in Deep Learning-Based Prediction of Type 1 Diabetic Retinopathy

Jui-En Lo et al. J Diabetes Res. .

Abstract

This study is aimed at evaluating a deep transfer learning-based model for identifying diabetic retinopathy (DR) that was trained using a dataset with high variability and predominant type 2 diabetes (T2D) and comparing model performance with that in patients with type 1 diabetes (T1D). The Kaggle dataset, which is a publicly available dataset, was divided into training and testing Kaggle datasets. In the comparison dataset, we collected retinal fundus images of T1D patients at Chang Gung Memorial Hospital in Taiwan from 2013 to 2020, and the images were divided into training and testing T1D datasets. The model was developed using 4 different convolutional neural networks (Inception-V3, DenseNet-121, VGG1, and Xception). The model performance in predicting DR was evaluated using testing images from each dataset, and area under the curve (AUC), sensitivity, and specificity were calculated. The model trained using the Kaggle dataset had an average (range) AUC of 0.74 (0.03) and 0.87 (0.01) in the testing Kaggle and T1D datasets, respectively. The model trained using the T1D dataset had an AUC of 0.88 (0.03), which decreased to 0.57 (0.02) in the testing Kaggle dataset. Heatmaps showed that the model focused on retinal hemorrhage, vessels, and exudation to predict DR. In wrong prediction images, artifacts and low-image quality affected model performance. The model developed with the high variability and T2D predominant dataset could be applied to T1D patients. Dataset homogeneity could affect the performance, trainability, and generalization of the model.

PubMed Disclaimer

Conflict of interest statement

The authors declare that there is no conflict of interest regarding the publication of this paper.

Figures

Figure 1
Figure 1
Fundus image after (a) cropping and (b) normalization.
Figure 2
Figure 2
Schematic of the development and evaluation of models. Two groups of models were trained using the T1D and Kaggle training sets, and they were tested with both the T1D and Kaggle testing sets.
Figure 3
Figure 3
Receiver operating characteristic (ROC) curves of different transfer learning models in predicting diabetic retinopathy. The ROC curve of models that were tested with the type 1 diabetes (T1D) testing set was plotted in blue, whereas those tested with the Kaggle testing set were plotted in orange. The point on the ROC curve was the selected threshold. (e)–(h) There was a significant decrease in AUC when models previously trained with the T1D training set were tested with the Kaggle dataset. (i)–(l) The models that were previously trained with the Kaggle training set have a more robust performance when tested with the T1D testing set.
Figure 4
Figure 4
The images demonstrate the original (a) and superimposed Grad-CAM activation maps ((b)–(i)) of the selected diabetic retinopathy (DR) color fundus image. All models gave a true-positive prediction. There were some similarities in activation maps even in different transfer learning models trained with different datasets.
Figure 5
Figure 5
The images present the original (a) and superimposed Grad-CAM activation maps ((b)–(i)) of the selected normal color fundus image. All models gave a true-negative prediction. There was a high variation in the activation map when given a normal fundus image. Some models focus on the optic disc ((e) and (g)), whereas others highlight the retinal vessels ((b), (c), and (g)), or macular region ((d), (h), (f), and (i)).
Figure 6
Figure 6
Images in the Kaggle dataset with wrong prediction. (a) False-negative in an image with foggy view and retinal laser scar. (b) False-negative in an image with poor illumination. (c) False-negative in an image with reflective spots and shadows. (d) False-positive in an image with overexposure and halo. (e) False-positive in an image with underexposure and halo. (f) False-positive in an image with exudates caused by age-related macular degeneration.

References

    1. Fong D. S., Aiello L., Gardner T. W., et al. Retinopathy in diabetes. Diabetes Care . 2004;27(Supplement 1):S84–S87. doi: 10.2337/diacare.27.2007.S84. - DOI - PubMed
    1. Flaxel C. J., Adelman R. A., Bailey S. T., et al. Diabetic retinopathy preferred practice pattern®. Ophthalmology . 2020;127:66–145. - PubMed
    1. American Diabetes Association. Standards of medical care in diabetes. Diabetes Care . 2004;27(Supplement 1):S15–S35. - PubMed
    1. Chou C.-F., Sherrod C. E., Zhang X., et al. Barriers to eye care among people aged 40 years and older with diagnosed diabetes, 2006-2010. Diabetes Care . 2014;37:180–188. - PMC - PubMed
    1. Gulshan V., Peng L., Coram M., et al. Development and validation of a deep learning algorithm for detection of diabetic retinopathy in retinal fundus photographs. JAMA . 2016;316:2402–2410. - PubMed

MeSH terms