MR PROSTATE SEGMENTATION VIA DISTRIBUTED DISCRIMINATIVE DICTIONARY (DDD) LEARNING
- PMID: 25035792
- PMCID: PMC4097123
- DOI: 10.1109/ISBI.2013.6556613
MR PROSTATE SEGMENTATION VIA DISTRIBUTED DISCRIMINATIVE DICTIONARY (DDD) LEARNING
Abstract
Segmenting prostate from MR images is important yet challenging. Due to non-Gaussian distribution of prostate appearances in MR images, the popular active appearance model (AAM) has its limited performance. Although the newly developed sparse dictionary learning method[1, 2] can model the image appearance in a non-parametric fashion, the learned dictionaries still lack the discriminative power between prostate and non-prostate tissues, which is critical for accurate prostate segmentation. In this paper, we propose to integrate deformable model with a novel learning scheme, namely the Distributed Discriminative Dictionary (DDD) learning, which can capture image appearance in a non-parametric and discriminative fashion. In particular, three strategies are designed to boost the tissue discriminative power of DDD. First, minimum Redundancy Maximum Relevance (mRMR) feature selection is performed to constrain the dictionary learning in a discriminative feature space. Second, linear discriminant analysis (LDA) is employed to assemble residuals from different dictionaries for optimal separation between prostate and non-prostate tissues. Third, instead of learning the global dictionaries, we learn a set of local dictionaries for the local regions (each with small appearance variations) along prostate boundary, thus achieving better tissue differentiation locally. In the application stage, DDDs will provide the appearance cues to robustly drive the deformable model onto the prostate boundary. Experiments on 50 MR prostate images show that our method can yield a Dice Ratio of 88% compared to the manual segmentations, and have 7% improvement over the conventional AAM.
Keywords: Prostate segmentation; deformable segmentation; magnetic resonance image; sparse dictionary learning.
Figures





References
-
- Wright J, Yang AY, Ganesh A, et al. Robust face recognition via sparse representation. IEEE PAMI. 2009;31(no 2):210–227. - PubMed
-
- Zhang S, et al. Deformable segmentation via sparse representation and dictionary learning. MIA. 2012;16(no 7):1385–1396. - PubMed
-
- Hricak H, et al. Imaging Prostate Cancer: A Multidisciplinary Perspective1. Radiology. 2007 Apr;243(no 1):28–53. - PubMed
-
- Toth R, Madabhushi A. Multifeature landmark-free active appearance models: application to prostate MRI segmentation. IEEE TMI. 2012;31(no 8):1638–1650. - PubMed
-
- Kirschner M, Jung F, Wesarg S. Automatic prostate segmentation in MR images with a probabilistic active shape model. MICCAI PROMISE’12 Challenge. 2012:28–35.
Grants and funding
LinkOut - more resources
Other Literature Sources