A fully automatic unsupervised segmentation framework for the brain tissues in MR images
Paper in proceeding, 2014

This paper presents a novel fully automatic unsupervised framework for the segmentation of brain tissues in magnetic resonance (MR) images. The framework is a combination of our proposed Bayesian-based adaptive mean shift (BAMS), a priori spatial tissue probability maps and fuzzy c-means. BAMS is applied to cluster the tissues in the joint spatialintensity feature space and then a fuzzy c-means algorithm is employed with initialization by a priori spatial tissue probability maps to assign the clusters into three tissue types; white matter (WM), gray matter (GM) and cerebrospinal fluid (CSF). The proposed framework is validated on multimodal synthetic as well as on real T1-weighted MR data with varying noise characteristics and spatial intensity inhomogeneity. The performance of the proposed framework is evaluated relative to our previous method BAMS and other existing adaptive mean shift framework. Both of these are based on the mode pruning and voxel weighted k-means algorithm for classifying the clusters into WM, GM and CSF tissue. The experimental results demonstrate the robustness of the proposed framework to noise and spatial intensity inhomogeneity, and that it exhibits a higher degree of segmentation accuracy in segmenting both synthetic and real MR data compared to competing methods.

Brain segmentation

Spatial tissue probability map

Fuzzy c-means

Mean shift

Magnetic resonance

Author

Mahmood Qaiser

Chalmers, Signals and Systems, Signal Processing and Biomedical Engineering

Artur Chodorowski

Chalmers, Signals and Systems, Signal Processing and Biomedical Engineering

B. Ehteshami Bejnordi

Radboud University

Mikael Persson

Chalmers, Signals and Systems, Signal Processing and Biomedical Engineering

Progress in Biomedical Optics and Imaging - Proceedings of SPIE

16057422 (ISSN)

Vol. 9038 90381M
978-081949831-1 (ISBN)

Subject Categories

Medical Engineering

DOI

10.1117/12.2043646

ISBN

978-081949831-1

More information

Latest update

4/4/2018 8