Advancing Volumetric Medical Image Segmentation via Global-Local Masked Autoencoders
- PMID: 40366838
- DOI: 10.1109/TMI.2025.3569782
Advancing Volumetric Medical Image Segmentation via Global-Local Masked Autoencoders
Abstract
Masked Autoencoder (MAE) is a self-supervised pre-training technique that holds promise in improving the representation learning of neural networks. However, the current application of MAE directly to volumetric medical images poses two challenges: (i) insufficient global information for clinical context understanding of the holistic data, and (ii) the absence of any assurance of stabilizing the representations learned from randomly masked inputs. To conquer these limitations, we propose the Global-Local Masked AutoEncoders (GL-MAE), a simple yet effective selfsupervised pre-training strategy. GL-MAE acquires robust anatomical structure features by incorporating multilevel reconstruction from fine-grained local details to high-level global semantics. Furthermore, a complete global view serves as an anchor to direct anatomical semantic alignment and stabilize the learning process through global-to-global consistency learning and global-to-local consistency learning. Our fine-tuning results on eight mainstream public datasets demonstrate the superiority of our method over other state-of-the-art self-supervised algorithms, highlighting its effectiveness on versatile volumetric medical image segmentation and classification tasks.We will release codes upon acceptance at https://github.com/JiaxinZhuang/GL-MAE.
Similar articles
-
MiM: Mask in Mask Self-Supervised Pre-Training for 3D Medical Image Analysis.IEEE Trans Med Imaging. 2025 Apr 25;PP. doi: 10.1109/TMI.2025.3564382. Online ahead of print. IEEE Trans Med Imaging. 2025. PMID: 40279226
-
Swin MAE: Masked autoencoders for small datasets.Comput Biol Med. 2023 Jul;161:107037. doi: 10.1016/j.compbiomed.2023.107037. Epub 2023 May 23. Comput Biol Med. 2023. PMID: 37230020
-
GMIM: Self-supervised pre-training for 3D medical image segmentation with adaptive and hierarchical masked image modeling.Comput Biol Med. 2024 Jun;176:108547. doi: 10.1016/j.compbiomed.2024.108547. Epub 2024 May 6. Comput Biol Med. 2024. PMID: 38728994
-
GO-MAE: Self-supervised pre-training via masked autoencoder for OCT image classification of gynecology.Neural Netw. 2025 Jan;181:106817. doi: 10.1016/j.neunet.2024.106817. Epub 2024 Oct 18. Neural Netw. 2025. PMID: 39500244 Review.
-
Local contrastive loss with pseudo-label based self-training for semi-supervised medical image segmentation.Med Image Anal. 2023 Jul;87:102792. doi: 10.1016/j.media.2023.102792. Epub 2023 Mar 11. Med Image Anal. 2023. PMID: 37054649
LinkOut - more resources
Full Text Sources