Rethinking Domain-Specific Pretraining by Supervised or Self-Supervised Learning for Chest Radiograph Classification: A Comparative Study Against ImageNet Counterparts in Cold-Start Active Learning
- PMID: 40241982
- PMCID: PMC11997468
- DOI: 10.1002/hcs2.70009
Rethinking Domain-Specific Pretraining by Supervised or Self-Supervised Learning for Chest Radiograph Classification: A Comparative Study Against ImageNet Counterparts in Cold-Start Active Learning
Abstract
Objective: Deep learning (DL) has become the prevailing method in chest radiograph analysis, yet its performance heavily depends on large quantities of annotated images. To mitigate the cost, cold-start active learning (AL), comprising an initialization followed by subsequent learning, selects a small subset of informative data points for labeling. Recent advancements in pretrained models by supervised or self-supervised learning tailored to chest radiograph have shown broad applicability to diverse downstream tasks. However, their potential in cold-start AL remains unexplored.
Methods: To validate the efficacy of domain-specific pretraining, we compared two foundation models: supervised TXRV and self-supervised REMEDIS with their general domain counterparts pretrained on ImageNet. Model performance was evaluated at both initialization and subsequent learning stages on two diagnostic tasks: psychiatric pneumonia and COVID-19. For initialization, we assessed their integration with three strategies: diversity, uncertainty, and hybrid sampling. For subsequent learning, we focused on uncertainty sampling powered by different pretrained models. We also conducted statistical tests to compare the foundation models with ImageNet counterparts, investigate the relationship between initialization and subsequent learning, examine the performance of one-shot initialization against the full AL process, and investigate the influence of class balance in initialization samples on initialization and subsequent learning.
Results: First, domain-specific foundation models failed to outperform ImageNet counterparts in six out of eight experiments on informative sample selection. Both domain-specific and general pretrained models were unable to generate representations that could substitute for the original images as model inputs in seven of the eight scenarios. However, pretrained model-based initialization surpassed random sampling, the default approach in cold-start AL. Second, initialization performance was positively correlated with subsequent learning performance, highlighting the importance of initialization strategies. Third, one-shot initialization performed comparably to the full AL process, demonstrating the potential of reducing experts' repeated waiting during AL iterations. Last, a U-shaped correlation was observed between the class balance of initialization samples and model performance, suggesting that the class balance is more strongly associated with performance at middle budget levels than at low or high budgets.
Conclusions: In this study, we highlighted the limitations of medical pretraining compared to general pretraining in the context of cold-start AL. We also identified promising outcomes related to cold-start AL, including initialization based on pretrained models, the positive influence of initialization on subsequent learning, the potential for one-shot initialization, and the influence of class balance on middle-budget AL. Researchers are encouraged to improve medical pretraining for versatile DL foundations and explore novel AL methods.
Keywords: COVID‐19; chest radiograph analysis; cold‐start active learning; psychiatric pneumonia; radiology foundation model.
© 2025 The Author(s). Health Care Science published by John Wiley & Sons, Ltd on behalf of Tsinghua University Press.
Conflict of interest statement
The authors declare no conflicts of interest.
Figures





Similar articles
-
Weakly-supervised learning-based pathology detection and localization in 3D chest CT scans.Med Phys. 2024 Nov;51(11):8272-8282. doi: 10.1002/mp.17302. Epub 2024 Aug 14. Med Phys. 2024. PMID: 39140793
-
Why does my medical AI look at pictures of birds? Exploring the efficacy of transfer learning across domain boundaries.Comput Methods Programs Biomed. 2025 Apr;261:108634. doi: 10.1016/j.cmpb.2025.108634. Epub 2025 Jan 31. Comput Methods Programs Biomed. 2025. PMID: 39913993
-
Enhancing diagnostic deep learning via self-supervised pretraining on large-scale, unlabeled non-medical images.Eur Radiol Exp. 2024 Feb 8;8(1):10. doi: 10.1186/s41747-023-00411-3. Eur Radiol Exp. 2024. PMID: 38326501 Free PMC article.
-
A survey of the impact of self-supervised pretraining for diagnostic tasks in medical X-ray, CT, MRI, and ultrasound.BMC Med Imaging. 2024 Apr 6;24(1):79. doi: 10.1186/s12880-024-01253-0. BMC Med Imaging. 2024. PMID: 38580932 Free PMC article. Review.
-
AMMU: A survey of transformer-based biomedical pretrained language models.J Biomed Inform. 2022 Feb;126:103982. doi: 10.1016/j.jbi.2021.103982. Epub 2021 Dec 31. J Biomed Inform. 2022. PMID: 34974190 Review.
References
-
- Litjens G., Kooi T., Bejnordi B. E., et al., “A Survey on Deep Learning in Medical Image Analysis,” Medical Image Analysis 42 (2017): 60–88. - PubMed
-
- Çallı E., Sogancioglu E., van Ginneken B., van Leeuwen K. G., and Murphy K., “Deep Learning for Chest X‐Ray Analysis: A Survey,” Medical Image Analysis 72 (2021): 102125. - PubMed
-
- Yuan H., Yu K., Xie F., Liu M., and Sun S., “Automated Machine Learning With Interpretation: A Systematic Review of Methodologies and Applications in Healthcare,” Medicine Advances 2, no. 3 (2024): 205–237.
-
- Sayin B., Krivosheev E., Yang J., Passerini A., and Casati F., “A Review and Experimental Analysis of Active Learning Over Crowdsourced Data,” Artificial Intelligence Review 54 (2021): 5283–5305.
LinkOut - more resources
Full Text Sources