Intensive vision-guided network for radiology report generation
- PMID: 38157546
- DOI: 10.1088/1361-6560/ad1995
Intensive vision-guided network for radiology report generation
Abstract
Objective.Automatic radiology report generation is booming due to its huge application potential for the healthcare industry. However, existing computer vision and natural language processing approaches to tackle this problem are limited in two aspects. First, when extracting image features, most of them neglect multi-view reasoning in vision and model single-view structure of medical images, such as space-view or channel-view. However, clinicians rely on multi-view imaging information for comprehensive judgment in daily clinical diagnosis. Second, when generating reports, they overlook context reasoning with multi-modal information and focus on pure textual optimization utilizing retrieval-based methods. We aim to address these two issues by proposing a model that better simulates clinicians perspectives and generates more accurate reports.Approach.Given the above limitation in feature extraction, we propose a globally-intensive attention (GIA) module in the medical image encoder to simulate and integrate multi-view vision perception. GIA aims to learn three types of vision perception: depth view, space view, and pixel view. On the other hand, to address the above problem in report generation, we explore how to involve multi-modal signals to generate precisely matched reports, i.e. how to integrate previously predicted words with region-aware visual content in next word prediction. Specifically, we design a visual knowledge-guided decoder (VKGD), which can adaptively consider how much the model needs to rely on visual information and previously predicted text to assist next word prediction. Hence, our final intensive vision-guided network framework includes a GIA-guided visual encoder and the VKGD.Main results.Experiments on two commonly-used datasets IU X-RAY and MIMIC-CXR demonstrate the superior ability of our method compared with other state-of-the-art approaches.Significance.Our model explores the potential of simulating clinicians perspectives and automatically generates more accurate reports, which promotes the exploration of medical automation and intelligence.
Keywords: multimodal learning; radiology report generation; visual reasoning; x-ray images.
© 2024 Institute of Physics and Engineering in Medicine.
Similar articles
-
Translating medical image to radiological report: Adaptive multilevel multi-attention approach.Comput Methods Programs Biomed. 2022 Jun;221:106853. doi: 10.1016/j.cmpb.2022.106853. Epub 2022 May 4. Comput Methods Programs Biomed. 2022. PMID: 35561439
-
S3-Net: A Self-Supervised Dual-Stream Network for Radiology Report Generation.IEEE J Biomed Health Inform. 2024 Mar;28(3):1448-1459. doi: 10.1109/JBHI.2023.3345932. Epub 2024 Mar 6. IEEE J Biomed Health Inform. 2024. PMID: 38133975
-
Radiology report generation with a learned knowledge base and multi-modal alignment.Med Image Anal. 2023 May;86:102798. doi: 10.1016/j.media.2023.102798. Epub 2023 Mar 23. Med Image Anal. 2023. PMID: 36989850
-
A survey of deep-learning-based radiology report generation using multimodal inputs.Med Image Anal. 2025 Jul;103:103627. doi: 10.1016/j.media.2025.103627. Epub 2025 May 13. Med Image Anal. 2025. PMID: 40382855 Review.
-
Integrating language into medical visual recognition and reasoning: A survey.Med Image Anal. 2025 May;102:103514. doi: 10.1016/j.media.2025.103514. Epub 2025 Feb 27. Med Image Anal. 2025. PMID: 40023891 Review.
Cited by
-
A vision attention driven Language framework for medical report generation.Sci Rep. 2025 Mar 28;15(1):10704. doi: 10.1038/s41598-025-95666-8. Sci Rep. 2025. PMID: 40155699 Free PMC article.
MeSH terms
LinkOut - more resources
Full Text Sources