Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2023 Jul;10(4):044503.
doi: 10.1117/1.JMI.10.4.044503. Epub 2023 Aug 4.

Effect of image resolution on automated classification of chest X-rays

Affiliations

Effect of image resolution on automated classification of chest X-rays

Md Inzamam Ul Haque et al. J Med Imaging (Bellingham). 2023 Jul.

Abstract

Purpose: Deep learning (DL) models have received much attention lately for their ability to achieve expert-level performance on the accurate automated analysis of chest X-rays (CXRs). Recently available public CXR datasets include high resolution images, but state-of-the-art models are trained on reduced size images due to limitations on graphics processing unit memory and training time. As computing hardware continues to advance, it has become feasible to train deep convolutional neural networks on high-resolution images without sacrificing detail by downscaling. This study examines the effect of increased resolution on CXR classification performance.

Approach: We used the publicly available MIMIC-CXR-JPG dataset, comprising 377,110 high resolution CXR images for this study. We applied image downscaling from native resolution to 2048×2048 pixels, 1024×1024 pixels, 512×512 pixels, and 256×256 pixels and then we used the DenseNet121 and EfficientNet-B4 DL models to evaluate clinical task performance using these four downscaled image resolutions.

Results: We find that while some clinical findings are more reliably labeled using high resolutions, many other findings are actually labeled better using downscaled inputs. We qualitatively verify that tasks requiring a large receptive field are better suited to downscaled low resolution input images, by inspecting effective receptive fields and class activation maps of trained models. Finally, we show that stacking an ensemble across resolutions outperforms each individual learner at all input resolutions while providing interpretable scale weights, indicating that diverse information is extracted across resolutions.

Conclusions: This study suggests that instead of focusing solely on the finest image resolutions, multi-scale features should be emphasized for information extraction from high-resolution CXRs.

Keywords: chest X-ray; deep learning; image resolution; multitask classification; receptive field.

PubMed Disclaimer

Figures

Fig. 1
Fig. 1
Distribution of 2D image sizes in the MIMIC-CXR-JPG dataset. Marginal distributions are shown as histograms along the top and right axes. The scatter plot shows many actual native resolutions, but the marginal histograms show that these are concentrated in a few common resolutions: 2000×2000, 2500×2500, 2500×3000, and 3000×2500  pixels.
Fig. 2
Fig. 2
ERF of trained DenseNet121 models relative to image size, for different image resolutions. As the image resolution increases, ERF decreases due to corresponding decrease in pixel size.
Fig. 3
Fig. 3
Workflow for the stacked ensemble model for a single input image. First, the image is resized to 256×256, 512×512, 1024×1024, and 2048×2048 resolution images and then passed to their corresponding fine-tuned models. The outputs from the models are then passed to a sigmoid function to get the output probabilities for each model. Finally, for each of the 14 labels, the predictions from the 4 models are multiplied by the learned weights to get the stacked ensemble prediction.
Fig. 4
Fig. 4
(a) Image-level prediction where each image is treated separately and given a label corresponding to the radiology report of the study it belongs to. (b) Study-level prediction where images corresponding to a study are aggregated first and then given a label according to the radiology report of the study.
Fig. 5
Fig. 5
Grad-CAM computed to visualize influential regions for predicting “cardiomegaly” (top), “pneumonia” (middle), and “pneumothorax” (bottom). Ground truth labels are shown overlaid on the original image, whereas for each Grad-CAM image, labels with probabilities >50% are shown. The top row shows that “cardiomegaly” is only correctly predicted at 256×256 resolution, presumably due to ERFs at high resolution being too small to encompass the entire heart. The middle row is a perfect example of the effect of receptive field on resolution. As the resolution is increasing, the model is able to better predict “pneumonia.” But at resolution 2048×2048, the ERF becomes too small to predict “pneumonia.” Conversely, the bottom row shows unreliable prediction of “pneumothorax” at coarse resolution due to lack of fine-scale information.
Fig. 6
Fig. 6
Barplot showing average stacked DenseNet scale-ensemble weights for each resolution for each task. Tasks requiring mostly coarse-scale information, such as cardiomegaly, place larger weight on resolutions 256×256 and 512×512, whereas others such as support devices focus on higher resolutions.

References

    1. Candemir S., Antani S., “A review on lung boundary detection in chest X-rays,” Int. J. Comput. Assist. Radiol. Surg. 14(4), 563–576 (2019).10.1007/s11548-019-01917-1 - DOI - PMC - PubMed
    1. Rosenkrantz A. B., Hughes D. R., Duszak R., “The U.S. radiologist workforce: an analysis of temporal and geographic variation by using large national datasets,” Radiology 279(1), 175–184 (2016).RADLAX10.1148/radiol.2015150921 - DOI - PubMed
    1. Yates E. J., Yates L. C., Harvey H., “Machine learning “red dot”: open-source, cloud, deep convolutional neural networks in chest radiograph binary normality classification,” Clin. Radiol. 73(9), 827–831 (2018).10.1016/j.crad.2018.05.015 - DOI - PubMed
    1. Dunnmon J. A., et al. , “Assessment of convolutional neural networks for automated classification of chest radiographs,” Radiology 290(2), 537–544 (2019).RADLAX10.1148/radiol.2018181422 - DOI - PMC - PubMed
    1. LeCun Y., Bengio Y., Hinton G., “Deep learning,” Nature 521(7553), 436–444 (2015).10.1038/nature14539 - DOI - PubMed