Self-improving generative foundation model for synthetic medical image generation and clinical applications
- PMID: 39663467
- DOI: 10.1038/s41591-024-03359-y
Self-improving generative foundation model for synthetic medical image generation and clinical applications
Abstract
In many clinical and research settings, the scarcity of high-quality medical imaging datasets has hampered the potential of artificial intelligence (AI) clinical applications. This issue is particularly pronounced in less common conditions, underrepresented populations and emerging imaging modalities, where the availability of diverse and comprehensive datasets is often inadequate. To address this challenge, we introduce a unified medical image-text generative model called MINIM that is capable of synthesizing medical images of various organs across various imaging modalities based on textual instructions. Clinician evaluations and rigorous objective measurements validate the high quality of MINIM's synthetic images. MINIM exhibits an enhanced generative capability when presented with previously unseen data domains, demonstrating its potential as a generalist medical AI (GMAI). Our findings show that MINIM's synthetic images effectively augment existing datasets, boosting performance across multiple medical applications such as diagnostics, report generation and self-supervised learning. On average, MINIM enhances performance by 12% for ophthalmic, 15% for chest, 13% for brain and 17% for breast-related tasks. Furthermore, we demonstrate MINIM's potential clinical utility in the accurate prediction of HER2-positive breast cancer from MRI images. Using a large retrospective simulation analysis, we demonstrate MINIM's clinical potential by accurately identifying targeted therapy-sensitive EGFR mutations using lung cancer computed tomography images, which could potentially lead to improved 5-year survival rates. Although these results are promising, further validation and refinement in more diverse and prospective settings would greatly enhance the model's generalizability and robustness.
© 2024. The Author(s), under exclusive licence to Springer Nature America, Inc.
Conflict of interest statement
Competing interests: The authors declare no competing interests.
References
-
- Gao, Y., Baptista-Hon, D. T. & Zhang, K. The inevitable transformation of medicine and research by large language models: the possibilities and pitfalls. MedComm Futur. Med. 2, e49 (2023). - DOI
-
- Wang, D.-Q., Feng, L.-Y., Ye, J.-G., Zou, J.-G. & Zheng, Y.-F. Accelerating the integration of ChatGPT and other large-scale AI models into biomedical research and healthcare. MedComm Futur. Med. 2, e43 (2023).
-
- Xia, K. & Wang, J. Recent advances of transformers in medical image analysis: a comprehensive review. MedComm Futur. Med. 2, e38 (2023). - DOI
-
- Ye, Y., Sarkar, S., Bhaskar, A., Tomlinson, B. & Monteiro, O. Using ChatGPT in a clinical setting: a case report. MedComm Futur. Med. 2, e51 (2023). - DOI
MeSH terms
Substances
LinkOut - more resources
Full Text Sources
Medical
Research Materials
Miscellaneous