Computational and Mathematical Methods in Medicine

Computational and Mathematical Methods in Medicine / 2015 / Article
Special Issue

Computational Intelligence Techniques in Medicine

View this Special Issue

Review Article | Open Access

Volume 2015 |Article ID 450341 | 23 pages | https://doi.org/10.1155/2015/450341

MRI Segmentation of the Human Brain: Challenges, Methods, and Applications

Academic Editor: Rafael M. Luque-Baena
Received27 Jun 2014
Revised11 Sep 2014
Accepted01 Oct 2014
Published01 Mar 2015

Abstract

Image segmentation is one of the most important tasks in medical image analysis and is often the first and the most critical step in many clinical applications. In brain MRI analysis, image segmentation is commonly used for measuring and visualizing the brain’s anatomical structures, for analyzing brain changes, for delineating pathological regions, and for surgical planning and image-guided interventions. In the last few decades, various segmentation techniques of different accuracy and degree of complexity have been developed and reported in the literature. In this paper we review the most popular methods commonly used for brain MRI segmentation. We highlight differences between them and discuss their capabilities, advantages, and limitations. To address the complexity and challenges of the brain MRI segmentation problem, we first introduce the basic concepts of image segmentation. Then, we explain different MRI preprocessing steps including image registration, bias field correction, and removal of nonbrain tissue. Finally, after reviewing different brain MRI segmentation methods, we discuss the validation problem in brain MRI segmentation.

1. Introduction

Over the last few decades, the rapid development of noninvasive brain imaging technologies has opened new horizons in analysing and studying the brain anatomy and function. Enormous progress in accessing brain injury and exploring brain anatomy has been made using magnetic resonance imaging (MRI). The advances in brain MR imaging have also provided large amount of data with an increasingly high level of quality. The analysis of these large and complex MRI datasets has become a tedious and complex task for clinicians, who have to manually extract important information. This manual analysis is often time-consuming and prone to errors due to various inter- or intraoperator variability studies. These difficulties in brain MRI data analysis required inventions in computerized methods to improve disease diagnosis and testing. Nowadays, computerized methods for MR image segmentation, registration, and visualization have been extensively used to assist doctors in qualitative diagnosis.

Brain MRI segmentation is an essential task in many clinical applications because it influences the outcome of the entire analysis. This is because different processing steps rely on accurate segmentation of anatomical regions. For example, MRI segmentation is commonly used for measuring and visualizing different brain structures, for delineating lesions, for analysing brain development, and for image-guided interventions and surgical planning. This diversity of image processing applications has led to development of various segmentation techniques of different accuracy and degree of complexity.

In this paper we review the most popular methods commonly used for brain MRI segmentation. We highlight differences between them and discuss their capabilities, advantages, and limitations. To introduce the reader to the complexity of the brain MRI segmentation problem and address its challenges, we first introduce the basic concepts of image segmentation. This includes defining 2D and 3D images, describing an image segmentation problem and image features, and introducing MRI intensity distributions of the brain tissue. Then, we explain different MRI preprocessing steps including image registration, bias field correction, and removal of nonbrain tissue. Finally, after reviewing different brain MRI segmentation methods, we discuss the validation problem in brain MRI segmentation.

2. Basic Concepts

2.1. and 3D Images

An image can be defined as a function in 2D space or in 3D space, where , , and denote spatial coordinates. The values (or amplitudes) of the functions and are intensity values and are typically represented by a gray value in MRI of the brain; see Figure 1. Every image consists of a finite set of image elements called pixels in 2D space or voxels in 3D space. Each image element is uniquely specified by its intensity value and its coordinates for pixels and for voxels, where is the image row number, is the image column number, and is the slice number in a volumetric stack; see Figure 2.

To each image element is assigned a single value based on the average magnetic resonance characteristics present in the tissue corresponding to that element. The size of the element determines the spatial resolution, or the fineness of detail that can be distinguished in an image. Voxel/pixel sizes vary depending on imaging parameters, magnet strength, the time allowed for acquisition, and other factors, but often in standard MRI studies voxel sizes are on the order of 1-2 mm. Greater spatial resolution can be obtained with a longer scanning time, but this must be weighed against patient discomfort. In adult brain MRI studies image acquisition time is around 20 min, while in pediatric MRI studies image acquisition time is limited to between 5 and 15 min.

2.2. Image Segmentation

The goal of image segmentation is to divide an image into a set of semantically meaningful, homogeneous, and nonoverlapping regions of similar attributes such as intensity, depth, color, or texture. The segmentation result is either an image of labels identifying each homogeneous region or a set of contours which describe the region boundaries.

Fundamental components of structural brain MRI analysis include the classification of MRI data into specific tissue types and the identification and description of specific anatomical structures. Classification means to assign to each element in the image a tissue class, where the classes are defined in advance. The problems of segmentation and classification are interlinked because segmentation implies a classification, while a classifier implicitly segments an image. In the case of brain MRI, image elements are typically classified into three main tissue types: white matter (WM), gray matter (GM), and cerebrospinal fluid (CSF); see Figure 3. The segmentation results are further used in different applications such as for analyzing anatomical structures, for studying pathological regions, for surgical planning, and for visualization.

Image segmentation can be performed on 2D images, sequences of 2D images, or 3D volumetric imagery. Most of the image segmentation research has focused on 2D images. If the data is defined in 3D space (e.g., obtained from a series of MRI images), then typically each image “slice” is segmented individually in a “slice-by-slice” manner. This type of segmenting 3D image volumes often requires a postprocessing step to connect segmented 2D slices into a 3D volume or a continuous surface. Furthermore, the resulting segmentation can contain inconsistencies and nonsmooth surface due to omitting important anatomical information in 3D space. Therefore, the development of 3D segmentation algorithms is desired for more accurate segmentation of volumetric imagery. The main difference between 2D and 3D image segmentation is in the processing elements, pixels/voxels, respectively, and their 2D or 3D neighborhoods (see Section 2.3) over which image features are calculated (see Section 2.4). In practice, 2D image segmentation methods can be extended to 3D space, but often with the cost of an increased complexity of the method and slower computational time.

2.3. Modeling the Spatial Context

The use of spatial context or neighborhood information is of great importance in brain MRI segmentation. Unless the image is simply random noise, the intensity of an image pixel/voxel is highly statistically dependent on the gray intensities of its neighbors (surrounding pixels/voxels). Markov random field (MRF) theory provides a basis for modeling local properties of an image, where the global image properties follow the local interactions. MRF models have been successfully integrated in various brain MRI segmentation methods to decrease misclassification errors due to image noise [13].

First, let us introduce some notations. As has been described in Section 2.1, every pixel (or voxel) in an image can be represented with one node in the lattice . Let represent an intensity value of a single pixel (or voxel) with a position in an image defined over a finite lattice , where is the total number of image elements ( for a 2D image and for a 3D image). Let denote a neighboring system for a lattice , where represent a small neighborhood around , not including .

The nodes (pixels/voxels) in a lattice are related to one another via neighborhood system that can be defined as The neighboring relationship has the following properties: (i)a node does not belong to its own neighborhood: ;(ii)the neighboring relationship is mutual:

The set of neighbors of can be defined as the set of surrounding nodes within a radius of from the center : where is the Euclidean distance between neighboring pixels and and is an integer number.

The first and the second order neighborhoods are the most commonly used neighborhoods in image segmentation. The first order neighborhood consists of 4 nearest nodes in a 2D image and 6 nearest nodes in a 3D image, while the second order neighborhood consists of 8 nearest nodes in a 2D image and 18 nearest nodes in a 3D image; see Figure 4.

Markov random field model can be represented with a graph , where represents the nodes and determines the links (also called edges) that connect the nodes according to the neighborhood relationship. Such graph structure corresponds to an image, where nodes correspond to pixels (or voxels) and the links connecting the nodes represent the contextual dependency between pixels (or voxels). More reading about MRF can be found in [4].

2.4. Image Features

Image features represent distinctive characteristics of an object or an image structure to be segmented. Features rely on numerical measurements, including quantitative visual appearance and shape descriptors, that can help to discriminate between the structures of interest and their background. The outcome of image segmentation highly depends on appropriate feature selection (choosing the most relevant features) and accurate feature extraction.

Typically, statistical approach is used for feature extraction and classification in MRI, where pattern/texture is defined by a set of statistically extracted features represented as a vector in multidimensional feature space. The statistical features are based on first and second order statistics of gray level intensities in an image. First order features are derived from the image grey value histogram and include the intensity, mean, median, and standard deviation of the pixel values. Since these features do not incorporate any information on the spatial distribution of the pixel values, they are often used in combination with second order features. Second order descriptors are used to describe image texture and are typically computed using gray level cooccurrence matrix [5]. First and second order features are often called appearance features in the literature. This is because the visual appearance of an object of interest is typically associated with its pixel or voxel intensities (gray values in brain MRI) and spatial interaction between intensities (intensity cooccurrence) in an image.

Image segmentation based on individual pixel/voxel intensities (first order features) is feasible only when intensities of an object of interest and its background differ to a large extent. Then, the complete object or the majority of its pixels/voxels can be separated from the background by simply comparing the intensity values to the threshold (the intensity value that clearly separates the object from the background). The threshold is derived from the overall intensity distribution of the image. In the presence of image noise and other imaging artifacts, first order features are not sufficient for accurate brain MRI segmentation. In this case more powerful second order discriminative features have to be used that include spatial interaction between intensities. For instance, the appearance of tumour lesion in brain MRI can be associated with spatial patterns of local pixel/voxel intensity variations or empirical probability distributions of intensity cooccurrences. In the spatial interaction models each intensity depends on a subset of the neighboring intensities; see Figures 5 and 4. The most popular models that can capture local spatial interactions between pixels/voxels intensities are MRF models [4].

Additionally, image segmentation performance can be also improved by incorporating probabilistic prior shape models, which have been extensively used in medical image segmentation [610]. The probabilistic prior shape models specify an average shape and variation of an object of interest and are typically estimated from a population of coaligned images of the object (training data sets) [11].

One of the most popular features for image segmentation is edges. Edges refer to boundaries of an object surface where the intensities change sharply [12]. Such changes are typically detected by thresholding the first and second order spatial derivatives of the intensities (the intensity gradient and Laplacian). However, edges detected in this way are sensitive to image noise [13] and often require image smoothing as a preprocessing step [14, 15].

Another more robust method for edge detected is the phase congruency method [16, 17], which is a frequency-based method for feature detection. This feature detection method, using local phase and energy, is based on a plausible model of how mammalians detect edges suggested by Morrone and Owens [18] and successfully explains the psychophysical effect of human feature perception. Instead of searching the pixels/voxels in the image with sharp intensity changes, features such as step edges, lines, and corners are detected at points where the Fourier components of the image are maximally in phase.

2.5. Intensity Distribution in Brain MRI

The intensity of brain tissue is one of the most important features for brain MRI segmentation. However, when intensity values are corrupted by MRI artifacts such as image noise, partial volume effect (PVE), and bias field effect, intensity-based segmentation algorithms will lead to wrong results. Thus, to obtain relevant and accurate segmentation results, very often several preprocessing steps are necessary to prepare MRI data. For instance, it is necessary to remove background voxels, extract brain tissue, perform image registration for multimodal segmentation, and remove the bias field effect; see Figure 6.

In the case when the bias field, nonbrain structures (e.g., the skull and the scalp) and background voxels are removed, the histogram of the adult brain MRI has three main peaks corresponding to the three main tissue classes; see Figure 8(a). In the healthy adult brain, the intensity variation within tissue is small and the intensities inside the brain can be considered to be a piecewise constant intensity function, corrupted by noise and PVE. The PVE describes the loss of small tissue regions due to the limited resolution of the MRI scanner. It means that one pixel/voxel lies in the interface between two (or more) classes and is a mix of different tissues. This problem is even more critical in imaging of the small neonatal brain. The correction of PVE will be addressed in Section 4.6.

It has been shown that the noise in the magnitude images is governed by a Rician distribution, based on the assumption that the noise on the real and imaginary channels is Gaussian [19]. The probability density function for a Rician distribution is defined as where is the measured pixel/voxel intensity, is the image pixel/voxel intensity in the absence of noise, is the standard deviation of the Gaussian noise in the real and the imaginary images, and is the zero-order modified Bessel function of the first kind. The Rician probability density function (PDF) is plotted in Figure 7(a) for several values of the signal-to-noise ratio (SNR), where the SNR is defined as (the power ratio between the signal and the background noise).

A special case of the Rician distribution is in image regions where only noise is present and SNR (e.g., in the dark background areas of an MRI where no NMR signal is present). This special case of the Rician distribution where and is also known as the Rayleigh distribution:

In the image regions where the NMR signal is present and SNR , the noise distribution approximates a Gaussian distribution; see Figure 7. Thus, the problem of Rician noise in the brain MRI is often simplified in practice by assuming the Gaussian distribution for the noise: where , , and are the intensity, the standard deviation, and the mean value, respectively. Due to this approximation, the histogram of a bias-corrected brain MRI in the presence of noise can be described with a Gaussian mixture model (GMM), where each tissue class (WM, GM, and CSF) is modeled by a Gaussian distribution. However, in the presence of partial volume effects the tissue intensity distributions slightly diverge from a Gaussian distribution, as can be seen from the histogram in Figure 8(a) where histograms of the tissue classes are based on manual segmentation. Recently, the -stable distribution mixture model is also suggested as an alternative to the Gaussian mixture model to model the histogram of MRI data for more complex MRI segmentation [20]. Note that the -stable distribution is a generalization of the Gaussian distribution.

The MRI intensity distribution of the neonatal brain is more complex because the intensity variability within tissue cannot be neglected due to the process of myelination. The histogram of 1.5 T -W MRI of the neonatal brain is shown in Figure 8(b). The difference between the neonatal and the adult brain histogram is the existence of the myelinated and nonmyelinated WM in neonates, which are separated with GM intensities. Since nonmyelinated WM is more dominant than myelinated WM, -W MRI shows inverted WM/GM intensities in neonates in comparison to adults.

2.5.1. T1-W and T2-W Intensity Distribution

It can be noted from the 1D histogram of the bias-corrected -W MRI of an adult brain in Figure 8(a) that there is an overlap between different tissue classes. Also, it can be seen that an overlap between WM and GM tissue is higher than between GM and CSF. This overlap between the class distributions can cause ambiguities in the decision boundaries when intensity-based segmentation methods are used [21]. However, many researchers showed that adding additional MRI sequences with different contrast properties (e.g., T2-W MRI, Proton Density MRI) can improve intensity-based segmentation and help separate the class distributions [2224]; see Figure 13.

3. MRI Preprocessing

After MRI acquisition several preprocessing steps are necessary to prepare MR images for segmentation; see Figure 6. The most important steps include MRI bias field correction, image registration (in the case of multimodal image analysis), and removal of nonbrain tissue (also called a brain extraction).

3.1. Bias Field Correction

The bias field, also called the intensity inhomogeneity, is a low-frequency spatially varying MRI artifact causing a smooth signal intensity variation within tissue of the same physical properties; see Figure 6. The bias field arises from spatial inhomogeneity of the magnetic field, variations in the sensitivity of the reception coil, and the interaction between the magnetic field and the human body [25, 26]. The bias field is dependent on the strength of the magnetic field. When MR images are scanned at 0.5 T, the bias field is almost invisible and can be neglected. However, when MR images are acquired with modern high-field MR scanners with a magnetic field strength of 1.5 T, 3 T, or higher, the bias field is strong enough to cause problems and considerably affect MRI analysis. In practice, trained medical experts can make visual MRI analysis to certain levels of intensity inhomogeneity (10%–30%) [26]. In contrast, the performance of automatic MRI analysis and intensity-based segmentation methods decreases greatly in the presence of the bias field; see Figure 9. This is because most of the segmentation algorithms assume intensity homogeneity within each class. Therefore, the correction of the bias field is an important step for the efficient segmentation and registration of brain MRI.

The bias field is typically modeled as low-frequency multiplicative field [26, 27]. Suppose that we place all image elements , , , and , into an × 1 column vector , where , , represents the observed intensity of the th voxel and is the total number of image elements. The degradation effect of each image voxel can be expressed as where is an ideal intensity of the th voxel and is an unknown smoothly varying bias field. The problem of eliminating the bias field is the task of estimating .

If the intensities of MRI are logarithmically transformed, the multiplicative bias field becomes an additive bias field as follows:

This simplified multiplicative model is used in most state-of-the-art bias correction methods to represent the bias field [2628]. However, in reality there are certain limitations to the correctness of this model. Even though the model is consistent with the variations arising from the sensitivity of the receiver coil, the relationship between the measured and true intensities in MRI is more complicated. This is due to nonuniformity of the induced currents and spatial inhomogeneity of the excitation field, which depends on the geometry and electromagnetic properties of the subject as well as the coil polarization and pulse sequence [26]. In spite of these difficulties, the multiplicative low-frequency model is successfully used in practice to model the intensity inhomogeneity in brain MRI.

In the literature, various methods have been proposed to correct the bias field in MRI. One of the earliest methods proposed to correct the bias field is based on the manual labeling of the brain tissue voxels, which are then used to reconstruct the bias field in form of a parametric surface. The main disadvantage of this surface fitting method is the need for manual interaction. The bias field can be also estimated and corrected by using low-pass filtering [29], but this approach can introduce additional artifacts in the image because it also removes the low-frequency component of the true image data. Both the surface fitting method and the low-pass method can be improved and made fully automatic if they are coupled with automatic segmentation of the brain [30, 31]. Other approaches for the bias field correction include minimizing the image entropy [32], fitting the histogram of the local neighbourhood to global histogram of the image [28], maximizing the high-frequency content of the image [26], and using a registered template image [27].

The template is an image/volume which encodes the average probability of finding different kinds of tissues at each spatial location. The anatomical template is obtained by normalizing, aligning, and averaging of anatomical images from several different subjects. All the images are normalized in a standard stereotaxic space such as the Montreal Neurological Institute (MNI space) [33]. MNI is widely used to provide a common reference for the 3D localization of functional activation foci and anatomical structures, enabling the comparison of results obtained across different studies. The standard probabilistic atlas of the human brain consists of a template and three tissue probability maps for WM, GM, and CSF [33]. The tissue probability maps are obtained by normalizing and averaging a number of segmented subjects. The probabilistic atlas then describes the anatomical variability of the brain.

Image registration is a necessary step for the inclusion of probabilistic atlases as a prior knowledge of the brain anatomy into the segmentation method. A probabilistic atlas is often used to initialize and constrain the segmentation process. The prior knowledge of the brain anatomical structures can increase the robustness and accuracy of a segmentation method; see Section 4.3.

3.2. Image Registration

Image registration is the process of overlaying (spatially aligning) two or more images of the same content taken at different times, from different viewpoints, and/or by different sensors. Registration is required in medical image analysis for obtaining more complete information about the patient’s health when using multimodal images (e.g., MRI, CT, PET, and SPECT) and for treatment verification by comparison of pre- and postintervention images. In medical image registration the term coregistration is used for intrasubject registration (the alignment multimodal images of the same subject), realignment is used for motion correction within the same subject, and normalization is used for intersubject registration when several population groups are studied.

Image registration involves finding the transformation between images so that corresponding image features are spatially aligned. The spatial alignment is typically initialized using rigid or affine transformation [34]. A rigid transformation is a 6-parameter transformation composed of translation and rotation. If scaling and skewing are allowed, we obtain a 12-parameter affine transformation. A rigid registration is sufficient for intrasubject registration if the object of interest does not deform. This is a reasonable assumption for images of the brain if these are acquired at the same stage of brain development. However, if the task is to match images belonging to either different subjects (intersubject registration) or the same subject at different stages of brain development (e.g., growth in children, changes related to ageing, or atrophy due to disease), a nonrigid registration of the images is required to obtain satisfactory results. The nonrigid registration algorithms are typically based on either physical models for transformation such as elastic [35] or fluid deformation models [36] or a linear combination of smooth basis functions [37] or free-form deformations [38]. However, the problems in intersubject brain MRI registration will arise when brains include lesions or diseases, because it is not possible to match the same structures between healthy and diseased brains. A general review of registration techniques can be found in [3941].

3.3. Removal of Nonbrain Tissue

Nonbrain tissues such as fat, skull, or neck have intensities overlapping with intensities of brain tissues. Therefore, the brain has to be extracted before brain segmentation methods can be used. This step classifies voxels as brain or nonbrain. The result can be either a new image with just brain voxels or a binary mask, which has a value of 1 for brain voxels and 0 for the rest of tissues. In general, the brain voxels comprise GM, WM, and CSF of the cerebral cortex and subcortical structures, including the brain stem and cerebellum. The scalp, dura matter, fat, skin, muscles, eyes, and bones are always classified as nonbrain voxels.

The common method for brain extraction is to use prior information of the brain anatomy. A deformable template can be registered with an image and nonbrain tissue is then removed by transferring the brain mask from the template [42]. However, brain extraction using a probabilistic atlas is usually not very accurate and can cause misclassification around the brain boundary. An alternative method for extracting the brain is the brain extraction tool (BET) [43, 44], which is part of the publicly available software package FSL. This method finds the center of gravity of the brain and then inflates a sphere until the brain boundary is found. It has been proven to work in practice on good quality -W and -W images of the adult brain. An example of the brain extraction is shown in Figure 10.

4. MRI Segmentation Methods

In general, MRI segmentation is not a trivial task, because acquired MR images are imperfect and are often corrupted by noise and other image artifacts. The diversity of image processing applications has led to development of various techniques for image segmentation [4554]. This is because there is no single method that can be suitable for all images, nor are all methods equally good for a particular type of image. For example, some of the methods use only the gray level histogram, while some integrate spatial image information to be robust for noisy environments. Some methods use probabilistic or fuzzy set theoretic approaches, while some additionally integrate prior knowledge (specific image formation model, e.g., MRI brain atlas) to further improve segmentation performance.

However, most of the segmentation methods developed for one class of images can be easily applied/extended to another class of images. For example, the theory of graph cuts, although firstly developed for binary images [55], can be modified and used for MRI segmentation of the brain tissue. Also, unsupervised fuzzy clustering [45, 56, 57] has been successfully applied in different areas such as remote sensing, geology, and medical, biological, and molecular imaging.

The segmentation methods, with application to brain MRI, may be grouped as follows:(i)manual segmentation;(ii)intensity-based methods (including thresholding, region growing, classification, and clustering);(iii)atlas-based methods;(iv)surface-based methods (including active contours and surfaces, and multiphase active contours);(v)hybrid segmentation methods.

4.1. Manual Segmentation

Manual segmentation refers to the process where a human operator (e.g., expert physician) segments and labels an image by hand. This segmentation is typically done in a “slice-by-slice” manner for 3D volumetric imagery. The manual method is believed to be the most accurate because of the difficulty to accurately and reliably delineate structures in medical images. The segmentation difficulties are related to image quality and artifacts.

Given the improvements achieved over the past years by imaging tools (e.g., MR scanners resolve images at millimetric resolution), the manual segmentation has become an intensive and time-consuming task. A trained operator typically has to go through around eighty images, slice by slice, to extract the contours of the target structures. This manual segmentation is not only tedious but also particularly prone to errors, as assessed by various intra- or interoperator variability studies [58, 59]. Also, manual segmentation results are often difficult and even impossible to reproduce, because even experienced operators show significant variability with respect to their own previous delineation.

However, manual segmentation is still intensively used for defining a surrogate for true delineation (called “ground truth”) and quantitative evaluation of automated segmentation methods. Also, manual segmentation of different brain structures is a fundamental step in brain atlas formation and is used in atlas-based segmentation approaches [50, 51, 60].

For manual delineation, editing tools such as ITK-SNAP [61, 62] usually display 3D data in the form of 3 synchronized 2D orthogonal views (sagittal, coronal, and axial) onto which the operator draws the contour of the target structure. The output data therefore consists of a series of 2D contours from which a continuous 3D surface has to be extracted. This is a nontrivial postprocessing task and is prone to errors. For instance, due to interslice inconsistencies in segmentation, bumps in the reconstructed 3D surface are inevitable. More robust segmentation methods can usually be derived from true 3D structure models in that they can ensure globally smoother and more coherent surfaces across slices.

4.2. Intensity-Based Methods

Intensity-based segmentation methods classify individual pixels/voxels based on their intensity. In the case of the brain MRI, three main tissue classes, WM, GM, and CSF, can be distinguished based on intensity; see Figure 3. A more detailed classification is not possible because the intensity profiles of more detailed brain structures overlap. Even separation of the three main tissue classes based on intensity itself requires incorporating tools for dealing with artifacts in MRI, such as intensity inhomogeneity, noise, and partial volume, as well as overlap in intensities of brain and nonbrain tissue (e.g., the scalp has the same intensities as brain tissues).

Several intensity-based techniques are available for tissue classification. The most common method is the use of intensity histogram of all of the voxels and fitting Gaussian functions to the distribution. The probability of a given intensity corresponding to a given type of tissue can thus be inferred and voxels are assigned to tissue types accordingly. Additionally incorporating neighbourhood information helps to give preference to spatially homogeneous regions in the resulting segmentation. This can significantly decrease misclassification due to random noise in the image [1]. Additionally, probabilistic atlases can be included in the classification to inform whether a given location in the brain is likely to contain WM, GM, or CSF voxels [50].

4.2.1. Thresholding

Thresholding is the simplest image segmentation method. A thresholding procedure uses the intensity histogram and attempts to determine intensity values, called thresholds , which separates the desired classes. The segmentation is then achieved by grouping all pixels between thresholds into one class; see Figure 11. The thresholding methods have many variations: global (single threshold) or local threshold (depending on the position in the image), multithresholding, adaptive thresholding, and so forth. In the case of a single global threshold, segmentation of an image is defined as where is a segmented (thresholded) image, where pixels labeled with 1 correspond to object and pixels labeled with 0 correspond to background; see Figure 11(a).

Thresholding is fast and computationally efficient method but does not take into account the spatial characteristics of an image (neighborhood information). Thus thresholding is sensitive to noise and intensity inhomogeneities. In low-contrast images it tends to produce scattered groups of pixels rather than connected regions and requires connectivity algorithms as a postprocessing step.

In general, threshold-based segmentation methods are not suitable for textured images. This is because the perceptual qualities of textured images are based on higher order interactions between image elements or objects in the scene. However, in brain MRI segmentation, thresholding can be used to separate background voxels from the brain tissue or to initialize the tissue classes in iterative segmentation methods such as fuzzy -means clustering. A survey on thresholding techniques is provided in [63].

4.2.2. Region Growing

Region growing (also called region merging) is a technique for extracting a connected region of the image which consists of groups of pixels/voxels with similar intensities [64]. In its simplest form, region growing starts with a seed point (pixel/voxel) that belongs to the object of interest. The seed point can be manually selected by an operator or automatically initialised with a seed finding algorithm. Then, region growing examines all neighboring pixels/voxels and if their intensities are similar enough (satisfying a predefined uniformity or homogeneity criterion), they are added to the growing region. This procedure is repeated until no more pixels/voxels can be added to the region.

Region growing is suitable for segmentation of volumetric images which are composed of large connected homogeneous regions. Thus, it is successfully used in medical image analysis to segment different tissues, organs, or lesions from MR images. For example, it is used in brain MRI analysis for segmentation of brain vessels [65], brain tumour segmentation [66], or extraction of brain surface [67]. See an example of region growing segmentation in Figure 12.

The main disadvantage of the region growing method is its sensitivity to the initialization of seed point. By selecting a different seed point, the segmentation result can be completely different. If seed point and homogeneity criterion are not properly defined, the growing region can leak out and merge with the regions that do not belong to the object of interest. Also, region growing is sensitive to noise and segmented regions in the presence of noise can become disconnected or have holes. On the other hand, separate regions can become connected in the presence of partial volume effects.

4.2.3. Classification Methods

Classification methods use data with known labels to partition image feature space. Image features are typically intensity values but can be also related to texture or other image properties. Classification methods can be both supervised and unsupervised. Supervised classification requires training images, which are manually segmented and then used as references for automatic segmentation of new images. Next to the manual interaction that is laborious and time-consuming, another disadvantage of supervised classification methods is that they generally do not take into account the neighborhood information and thus they are sensitive to noise. Also, the use of the same training set for a large number of images can lead to biased results, which do not take into account anatomical and physiological variability between different subjects.

One of the simplest classifiers is the nearest-neighbor classifier [68], where each pixel/voxel is classified in the same class as the training datum with the closest intensity. A generalization of this approach is the -nearest-neighbor (NN) classifier, where the pixel/voxel is classified according to the majority vote of the closest training data. The NN classifier is considered a nonparametric classifier because it makes no underlying assumption about the statistical structure of the data. It is especially suitable if a large number of training data are available.

The NN classification method was applied to brain MRI segmentation by Warfield et al. [69]. In addition to image intensities, Warfield used spatial localization of brain structures (classes) in form of a nonrigidly registered template as an additional feature to enhance the classification process. The segmentation is then calculated in an iterative process by interleaving the segmentation refinement with updating the nonrigid alignment to the template. This procedure requires manual selection of a large number of training samples for each tissue class to train the NN classifier. Due to the manual interaction in the training phase, the method is not fully automatic and the results depend on particular choice of the training set. Cocosco et al. [70] developed a method for the robust selection of training samples to make the NN classification process fully automatic. This method is reported to deal well with anatomies which differ from the probabilistic atlas. However, it does not deal with the problem of natural intensity variation within each tissue class. Both methods require correction of the bias field as a preprocessing step.

One of the most commonly used parametric classifiers is the Bayesian classifier [30]. The Bayesian classifier models the probabilistic relationships between the attribute set and the class variables, which are then used for estimating the class probability of the unknown variable. This model involves Bayesian inference such as maximum a posteriori (MAP) estimation, where the goal is to estimate the label output image given the observed image by minimizing the posterior distribution of the possible labels : The Bayesian framework consists of three probability distributions: the prior distribution , the posterior distribution , and the conditional distribution (also called the likelihood). The prior distribution embodies the knowledge of likely configurations before an actual image is observed. The posterior distribution is derived after an observation has been made and the likelihood is defined as the probability of obtaining a particular observation given a set of model parameters.

The Bayes rule describes the relation between the posterior probability , prior probability , and likelihood as follows: Using definition (11), the MAP estimate can be written as where can be omitted because it is a constant in the case when is known. Since in many cases the probability distributions have exponential functions, this computation can be simplified by using a logarithmic transform:

In the case of the brain MRI segmentation, often it is assumed that the pixel intensities are independent samples from a mixture of Gaussian probability distributions. Training data is collected by obtaining representative samples from each component of the Gaussian mixture accordingly. Classification of new data is obtained by assigning each pixel to the class with the highest posterior probability.

Bayesian classifiers are used in the expectation-maximization (EM) segmentation methods which have been successfully implemented in several software packages used in the medical imaging community: SPM [2], FAST [3], FreeSurfer [21], and 3DSlicer [71]. All these methods implement segmentation and bias correction in the EM framework. They also include various additional improvements, such as nonrigid alignment of atlas [2], including neighbourhood information in the form of Markov random fields [3, 4] or using the -stable distribution mixture model as a generalization of the GMM [20]. More details about the Bayes’ theory can be found in [72].

4.2.4. Clustering Methods

Clustering methods are unsupervised segmentation methods that partition an image into clusters of pixels/voxels with similar intensities without using training images. In fact, clustering methods use the available image data to train themselves. The segmentation and training are done in parallel by iterating between two steps: data clustering and estimating the properties of each tissue class. The most commonly used clustering methods are the -means clustering [73], the fuzzy -means clustering [74, 75], and the expectation-maximisation (EM) method [1].

The -means clustering method partitions the input data into classes by iteratively computing a mean intensity for each class (also called centroid) and segmenting the image by classifying each pixel/voxel in the class with the closest centroid. The -means clustering is also known as a hard classification method because it forces each pixel/voxel to belong exclusively to one class in each iteration. The fuzzy -means clustering is soft classification method based on fuzzy set theory [76]. It is a generalization of the -means clustering because it allows each pixel/voxel to belong to multiple classes according to a certain membership value.

The FCM clustering algorithm is based on minimizing the following objective function: where is the number of image elements that need to be partitioned into clusters, is the membership function of the element (a feature vector at position ) belonging to the cluster, is the weighting exponent that controls the fuzziness of the resulting partition (most often is set to , if we have the -means clustering), and is the similarity measure between and the cluster center . The most commonly used similarity measure is the squared Euclidean distance .

The objective function (see (14)) is minimized under the following constraints: , , and . Considering these constraints and calculating the first derivatives of with respect to and and setting them to zero result in the following two conditions for minimizing :

The FCM algorithm iteratively optimizes , by evaluating (15), until the following stop criterion is satisfied: , where is the iteration index and is the norm. Once a membership value for each class is assigned to each pixel , defuzzification of the fuzzy clusters into its crisp version is done by assigning the pixel to the class with the highest membership value as follows:

The EM method is an iterative method for finding maximum likelihood or MAP estimates of a statistical model. It has the same soft classification principle as FCM method but typically assumes that MRI intensities of different brain tissues can be represented with a Gaussian mixture model. Even though clustering methods do not require training images, they do require some initial parameters and the EM method has shown the highest sensitivity to initialization in comparison to fuzzy -means and -means methods [1].

In general, the EM segmentation framework can be described as follows.

EM Approach for Brain MRI Segmentation. Firstly, initialize the EM algorithm. In the case of brain MRI segmentation, the GMM is used to initially estimate model parameters. Then, iterate between expectation step (E-step) and maximization step (M-step) until convergence.

E-Step. Estimate the brain tissue segmentation given the current estimate of model parameters. This step can include the use of neighbourhood information (e.g., in the form of MRF modeling).

M-Step. Estimate the model parameters. This step can consist of a combination of the following steps.(1)Estimate the intensity distribution parameters for each tissue class.(2)Estimate the bias correction parameters.(3)Estimate the registration parameters for alignment of probabilistic atlas with the image.

As it is the case with classification methods, clustering methods initially do not incorporate spatial neighborhood information and thus they are sensitive to noise and intensity inhomogeneities. To improve the clustering performance for images corrupted by noise, many extensions of the clustering algorithms have been proposed [49, 57, 7784]. The most common approach is to include feature information (e.g., intensity values) of the neighboring pixels into the modified FCM objective function [77, 79] or into a similarity measure between cluster centers and image elements [80]. Ahmed et al. [77] modified the objective function of the standard FCM algorithm to allow the immediate neighbours of the pixel to influence its labeling. Chen and Zhang [79] proposed two improvements of the Ahmed et al. algorithm to reduce the computational time. On the other hand, to keep the continuity from the FCM algorithm, Shen et al. [80] introduced a new similarity measure that depends on spatial neighbourhood information, where the degree of the neighbourhood attraction is optimized by a neural network. The clustering performance can also be enhanced by combining pixel-wise fuzzy classification with preprocessing (noise cleaning in the original image) [49, 78] and postprocessing (noise cleaning on the classified data) [78].

An example of the multimodal -W and -W MRI clustering of the adult brain is shown in Figure 13(a). In general, the shape of joints -W and -W MRI intensity distributions of different tissue classes depends on the image quality (the presence of noise, PVE, etc.). The shape of the classified data depends on the applied segmentation method. In the example in Figure 13(a), there is a small overlap among classes due to the good quality MRI. Thus, the standard -means clustering method is used to segment the brain tissue probability maps (see Section 4.2.4) and the final clusters are indicated with different colors in the scatter plot of -W and -W MRI in Figure 13(b). In general, when MRI artifacts are present and there is a significant overlap among tissue classes, the spatial information of the brain tissue is required to disambiguate the classification problem.

4.3. Atlas-Based Methods

If an atlas or template of the human brain for a specific population of interest is available, then atlas-based methods can be a powerful tool for brain MRI segmentation. The atlas contains information about the brain anatomy (e.g., it contains the information about the location of different brain structures) and it is used as a reference (a prior knowledge) for segmenting new images. The main advantage of these methods is the possibility to segment any brain structure available in the atlas without any additional cost. Conceptually, atlas-based approaches are similar to classifier methods, except that they are implemented in the spatial domain rather than in the feature space.

Before a probabilistic atlas can be used as a prior knowledge, it has to be aligned with the image to be segmented. Since the segmentation labels and the “ground truth” are known for the atlas, all atlas information is transferred to the target image after registration. Therefore, the performance of atlas-based methods is directly dependent on quality of the registration method used.

The traditional way of aligning the probabilistic atlas with the image is to use affine registration. Unfortunately, an affine alignment may not be sufficient if the brain anatomy of interest differs significantly from the average atlas anatomy. Pohl et al. therefore suggest aligning the atlas using nonrigid registration [85]. However, in their later work Pohl reports difficulties in registering anatomical template with the image to be segmented using standard registration methods [86]. D’Agostino et al. developed a special similarity measure for registering probabilistic maps directly to the new image [87]. Recently, several methods have been developed which aim to overcome this problem by iteratively refining the segmentation and nonrigid registration of the probabilistic atlas at the same time. Ashburner and Friston developed a method for simultaneous segmentation, bias correction, and nonrigid registration of a probabilistic atlas [2].

However, even with nonrigid registration methods, accurate segmentation of complex structures is difficult due to anatomical variability. Also, atlas-guided segmentation in patients with brain deformations can be difficult and prone to errors, because the probabilistic atlas is based on a population of healthy subjects. For instance, in patients with brain lesions or a brain anatomy that significantly differs from the atlas template, the atlas alignment and the corresponding segmentation of the brain will fail or give inaccurate results. In these cases an atlas-based approach is not a suitable method for image segmentation.

An aligned probabilistic atlas can be also used as a good initial estimate of the segmentation, which is especially important for EM-based methods, as EM algorithm is guaranteed to converge to local, not global, maxima. In addition, most EM-based methods [2, 71] use the probabilistic atlas to constrain the segmentation process where again the correct alignment of the probabilistic atlas is crucial for successful and accurate segmentation.

It is important to note that atlas-based MRI segmentation of the neonatal brain has become a research focus in recent years [42, 51, 53, 88]. MRI segmentation of the neonatal brain tissue is more complex than in adults due to fast growth process, complex anatomy of the developing brain, and often poor MRI quality. Therefore a probabilistic atlas of the newborn brain that contains the spatial variability of the tissue structure is used to segment different brain tissues such as brain cortex, myelinated and nonmyelinated white matter. However, a good atlas of the newborn brain is even more difficult to obtain than in adults, mainly due to the greater anatomical variations between subjects. Therefore, it was necessary to develop a dynamic, probabilistic atlas for any chosen stage of neonatal brain development (for ages of 29 to 44 weeks) [89].

4.4. Surface-Based Methods

In addition to intensity-based and atlas-based methods, there are a number of alternative brain MRI segmentation approaches. These approaches include surface-based methods, such as deformable models including active contours and surfaces [46, 52, 54, 61, 9093].

4.4.1. Active Contours and Surfaces

Deformable models are also called active contours or snakes in 2D and active surfaces or active balloons in 3D. Deformable models were introduced by Kass et al. [92] in 2D space and were further developed and generalized in 3D space by Terzopoulos et al. [93, 94]. Deformable models use closed parametric curves or surfaces for delineating region boundaries. The parametric curves and surfaces deform under the influence of external (or image) forces (controlled by the image attributes) and the internal forces, which control the surface regularity. In general, deformable models represent the fusion of geometry, physics, and approximation theory. Geometry is used to represent the shape of the object, physics defines constraints on how the shape may vary over time and space, and approximation theory provides mechanisms for fitting the models to measured data. A visual example of segmentation using deformable models is given in Figure 14.

To delineate a boundary of an object, first a closed curve or surface is placed near the desired boundary in an image. Then, internal and external forces are deforming the curve or surface in an iterative relaxation process where the energy functional is defined as The internal forces are computed from within the curve or surface to keep it smooth throughout the deformation. The external forces are usually derived from the image to deform the curve or surface towards the desired feature of interest.

In traditional deformable models, image forces come primarily from the local edge-based information (e.g., based on the gradients of sharp image intensities) [9597]. However, such reliance on edge information makes deformable models sensitive to noise (e.g., deformable model can leak through noisy edges) and highly dependent on the initial estimate. There have been significant efforts to integrate more global region information into deformable models. The Mumford-Shah model [98] was one of the first region-based methods where the image is approximated using a smooth function inside the regions and not only at their boundaries. Many variants of this model have been proposed later [46, 90]. For instance, Chan and Vese [46] presented a level set method which approximates an image with a constant function inside the regions. Li et al. proposed a region-based level set method for segmentation of MRI in the presence of intensity inhomogeneity. Furthermore, several hybrid deformable models had been later proposed to make use of both local (edge-based) and global (region-based) information [99102]. Sometimes the image data are not sufficient to delineate the region of interest and thus prior knowledge has to be introduced [103].

4.4.2. Multiphase Active Contours

The most popular Chan-Vese level set method [46] has been successfully used in segmenting images with two distinctive regions (images with binary segmentation energies). In [104], Vese and Chan extended their binary segmentation energies to a multiphase level set formulation. In this way, multiple nonoverlapping regions with spatial consistency and varying characteristics (such as the mean intensities of regions) could be represented with multiple level set functions. This multiphase level set approach was attractive for segmentation of brain MR images which typically have multiple regions of interest with different characteristics. Starting from the Vese and Chan method [104], different extension to multiphase active contours had been developed [54, 105108]. The advantage of multiphase active contours to other approaches is their robustness to image variations, adaptive energy functionals, topological flexibility, and accurate boundaries.

Traditionally, active contours methods have nonconvex energy minimization due to a gradient descent formulation. In this way, energy minimization converges to undesirable local minima and results in erroneous segmentations. Also, the traditional level set implementation has slower convergence due to discretization errors and the well-known reinitialization requirement. One of the first convex approaches to the two-phase active contours segmentation was proposed by Chan et al. [109]. Later on, several extensions to the more challenging multiphase problem have been proposed [110, 111]. A lot of recent work has been dedicated to developing new multiphase active contours not only with a convex formulation, but also with a reduced computational complexity [54, 105, 107]. Note that globally convex methods are initialization independent.

4.5. Hybrid Segmentation Methods

New application-specific brain MRI segmentation problems are emerging and new methods are continuously explored and introduced. Since a selection of the most appropriate technique for a given application is often a difficult task, a combination of several techniques may be necessary to obtain the segmentation goal. Therefore, hybrid or combined segmentation methods have been used extensively in different brain MRI segmentation applications [78, 84, 112120]. The main idea is to combine different complementary segmentation methods into a hybrid approach to avoid many of the disadvantages of each method alone and improve segmentation accuracy.

Here are some examples of the hybrid brain MRI segmentation methods. Kapur et al. [112] segmented different brain tissues in adults using 2D MRI by combining expectation-maximization segmentation, binary mathematical morphology, and active contours models. Masutani et al. [113] combined model-based region growing with morphological information of local shape to segment cerebral blood vessels. Warfield et al. [120] developed a combined 3D brain MRI segmentation algorithm which iterates between a classification step to identify tissues and an elastic matching step to align a template of normal brain anatomy with the classified tissues. Elastic matching step can generate image segmentation by registering an anatomical atlas to a patient scan.

Furthermore, an unsupervised global-to-local brain MRI segmentation is developed by Xue et al. [78]. They combined minimum error global thresholding and a spatial-feature-based FCM clustering to segment 3D MRI in a “slice-by-slice” manner. In the work of Vijayakumar and Gharpure [117], a hybrid MRI segmentation method, based on artificial neural networks (ANN), is proposed for segmenting tumor lesions, edema, cysts, necrosis, and normal tissue in T2 and FLAIR MRI. More recently, Ortiz et al. [119] suggested an improved brain MRI segmentation method using self-organizing maps (a particular case of ANN) and entropy-gradient clustering.

Hybrid segmentation methods are also used for the neonatal brain segmentation [115, 118]. For example, Despotovic et al. [115] proposed a hybrid strategy to segment the brain volume in neonates using -W and -W MRI by combining thresholding, active contours, FCM clustering, and morphological operations. Later on, Gui et al. [118] proposed a morphology-driven automatic segmentation method to segment different anatomical regions of the neonatal brain.

The main drawback of hybrid (combined) segmentation methods is often the increased complexity in comparison with each single method integrated into a hybrid one. This includes a lower computational time and a higher number of different parameters that needs to be tuned for a specific application. Therefore, a hybrid segmentation method should be carefully and wisely designed to give efficient and good quality segmentation.

4.6. Partial Volume Effect Correction

As mentioned in Section 2.5, the PVE problem is one of the most common problems in brain MRI segmentation. The PVE describes the loss of small tissue regions because of the limited resolution of the MRI scanner and it is seen on MRI scans as a mix of different tissues in a single pixel/voxel. This effect can cause the misclassification of pixel/voxel that lies in the transition between two (or more) tissues classes. Several methods have been proposed to address the problem of PVE in MRI segmentation for both adult and neonatal brains [3, 42, 103, 121, 122].

One of the first approaches for the partial volume correction is the method by Santago and Gage [123]. They assumed a uniform prior probability for mixed (nonpure) tissues and calculated the intensity distribution of partial volume by minimizing the distance between a model and an image histogram. Another approach is proposed by Nocera and Gee [124], where they used MRF to obtain spatial smooth variations of tissue mixing proportions and a MAP estimation is then computed for partial volume segmentation. Later on, Zhang et al. [3] employ a hidden Markov random field with a finite mixture model to overcome possible PVE and bias field distortions. In the work by Van Leemput et al. [122], a statistical uniform framework for partial volume segmentation is presented without using a heuristic assumption for the prior distribution of mix proportions. They used a parametric statistical image model where each voxel belongs to a single tissue type and introduced an additional downsampling step to cause partial volumes along the borders between tissues. Then, they estimated the tissue mixing proportions by using the expectation-maximization approach.

For all above-mentioned partial volume correction methods, very promising results have been reported for adult -W MRI. However, these methods rely on the fact that the intensity levels of partial volumes in adult MRI images do not predominantly overlap with the characteristic intensity of any pure tissue class. This assumption is not possible for neonatal MRI due to the inverted graywhite matter contrast and great tissue overlap due to the existence of myelinated WM and nonmyelinated WM; see Figure 8(b). Therefore, several approaches have been proposed to address the PVE in neonatal MRI [42, 103]. Xue et al. [42] proposed method based on expectation-maximization, Markov random field, and atlas information to remove mislabeled voxels and correct errors caused by PVE. They implemented a label propagation strategy to mask off deep GM and myelinated WM, which enabled the segmentation of cortical GM and nonmyelinated WM. In the paper of Wang et al. [103], a rather simple and effective scheme is used to deal with the PVE problem and is based on the anatomical observation that the misclassified WM voxels are surrounded by the CSF and GM and that mislabeled CSF voxels are unconnected from the true WM volume.

5. Validation of Brain MRI Segmentation

Validation and quantitative comparison of different segmentation methods are a general problem in medical image analysis. It requires a “ground truth” or gold standard to which the outcome of the segmentation method can be compared. Unfortunately, the “ground truth” does not exist for the analysis of in vivo acquired data in humans. Thus, the “ground truth” of the real patients is typically generated after image acquisition.

In brain MRI analysis, the “ground truth” for the real patient data is usually made by one or more expert physicians who need to manually analyze and segment anatomical structures of interest; see Section 4.1. Although this is the only way to validate the real patient MRI data, this validation must be critically considered because the manual segmentation is prone to errors, highly subjective, and difficult to reproduce (even by the same expert) [59]. Also, this type of validation is not always available because it is time-consuming and depends on the human operator. Therefore, few alternative validation methods evolved in the praxes to validate the accuracy of the segmentation algorithms. The most popular validation methods include the use of software simulations and phantoms.

In software simulations, the artificial MR images are generated with computer programs that simulate the real acquisition process. In this way the “ground truth” is known and the influence of different acquisition parameters and imaging artifacts can be controlled and examined independently. This type of validation is very flexible and easily accessible by different researchers and can be performed with little effort. However, a drawback of this validation is that software simulators cannot take into account all factors that might influence the real image acquisition and the simulated images are only an approximation of the real images.

Since software simulations have certain limitations, validation of new segmentation methods can be done using human-like phantoms, whose physical properties (e.g., geometry of the tissue structures and material properties) are known and are similar to the in vivo properties. The phantom images are generated using the MRI scanner and are more realistic than images generated with software simulations. On the other hand, the phantom images do not offer the flexibility of the software simulations and imaging is more expensive and labour intensive.

The most popular simulated images used for validation of brain MRI segmentation methods are designed by Collins et al. [125] and are also known as a realistic digital brain phantom or simply BrainWeb. Images are freely available online and easily accessible for all researchers to test the performance of the new segmentation methods. The BrainWeb data consists of 181 × 217 × 181 voxel matrix with a resolution of 1 mm ×  1 mm ×  1 mm and is available for different additive noise levels. The noise level (expressed in percentages) is relative to the average real and imaginary values of the overall brightness of the tissue class. The noise is generated using a pseudorandom Gaussian noise, which is added to both real and imaginary components before the final magnitude value of the simulated MR image is computed.

Beside the phantom BrainWeb data, the most popular repository with real MRI data used for validation of brain MRI segmentation methods is the Internet Brain Segmentation Repository (IBSR) [126]. The IBSR repository is also freely available online. It consists of 20 real -W MRI brain data sets and manually guided expert segmentation results, which are used as a “ground truth” segmentation. Each MRI volume consists of about 60 coronal -W slices with the interslice resolution of 3.1 mm (thickness between consecutive slices).

To quantify the overlap between the MRI segmentation and the given “ground truth,” several similarity measures are used in the literature. One of the most popular measures used often with BrainWeb data is the Dice coefficient [127]: where stands for a tissue type, and denote the set of pixels labeled into by the “ground truth” and MRI segmentation, respectively, and denotes the number of elements in . The Dice coefficient is in the range and has value 0 if there is no overlap between the two segmentations and 1 if both segmentations are identical.

The Tanimoto coefficient (also known as the Jaccard index) is often used as a similarity measure with real IBSR data. The Tanimoto coefficient for each tissue type is defined as follows: where and denote the set of pixels labeled into by the “ground truth” and the segmentation method, respectively, and denotes the number of elements in . Note that and .

More readings about similarity measures for evaluation and validation in medical image analysis can be found in [128].

6. Discussion and Conclusions

Image segmentation is an important step in many medical applications involving 3D visualization, computer-aided diagnosis, measurements, and registration. This paper has provided a brief introduction to the fundamental concepts of MRI segmentation of the human brain and methods that are commonly used.

In Section 2, we have defined the basic concepts necessary for understanding MRI segmentation methods, such as 2D and 3D image definition, image features, and brain MRI intensity distributions. Following this, preprocessing steps necessary to prepare images for MRI segmentation have been described in Section 3. The most important steps include bias field correction, image registration, and removal of nonbrain tissues or brain extraction. The correction of intensity inhomogeneity is an important step for the efficient segmentation and registration of brain MRI. Image registration is required in brain MRI segmentation for the alignment of multimodal images of the same subject or several population groups taken at different times and from different viewpoints.

Due to the rapid development of medical image modalities, new application-specific segmentation problems are emerging and new methods are continuously explored and introduced. Selection of the most appropriate technique for a given application is a difficult task. In many cases, a combination of several techniques may be necessary to obtain the segmentation goal. Very often integration of multimodal information (acquired from different modalities or over time) can help to segment structures that otherwise could not be detected on single images.

The most popular image segmentation methods that are used for brain MRI segmentation have been reviewed and discussed in Section 4. Newer methods are usually designed to bring more accurate results by incorporating 3D neighborhood information and prior information from atlases. As a consequence, the segmentation process often becomes more complex and time-consuming. The likely future research will still focus not only on developing more accurate and noise-robust methods, but also on improving the computational speed of segmentation methods. Computational efficiency will be particularly important in real-time processing applications such as computer guided surgery.

Probably one of the most important questions concerning medical image segmentation is its use in real clinical settings. It is undeniable that computerized segmentation methods have shown their potentials and applicability in computer-aided diagnosis and therapy planning. It is expected that in the near future they will also become essential tools in real clinical settings, particularly in qualitative diagnosis and where 3D reconstruction and visualization of the anatomical structures are important.

Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

Acknowledgments

This work was supported by FWO-Vlaanderen under Grant G.0341.07 “Data fusion of multimodal information using advanced signal processing, segmentation and registration techniques” and by IWT project NeoGuard.

References

  1. D. L. Pham, C. Xu, and J. L. Prince, “Current methods in medical image segmentation,” Annual Review of Biomedical Engineering, vol. 2, no. 2000, pp. 315–337, 2000. View at: Publisher Site | Google Scholar
  2. J. Ashburner and K. J. Friston, “Unified segmentation,” NeuroImage, vol. 26, no. 3, pp. 839–851, 2005. View at: Publisher Site | Google Scholar
  3. Y. Zhang, M. Brady, and S. Smith, “Segmentation of brain MR images through a hidden Markov random field model and the expectation-maximization algorithm,” IEEE Transactions on Medical Imaging, vol. 20, no. 1, pp. 45–57, 2001. View at: Publisher Site | Google Scholar
  4. S. Z. Li, Markov Random Field Modeling in Computer Vision, Springer, New York, NY, USA, 1995.
  5. R. M. Haralick, K. Shanmugam, and I. Dinstein, “Textural features for image classification,” IEEE Transactions on Systems, Man and Cybernetics, vol. 3, no. 6, pp. 610–621, 1973. View at: Publisher Site | Google Scholar
  6. P. Yan, S. Xu, B. Turkbey, and J. Kruecker, “Discrete deformable model guided by partial active shape model for trus image segmentation,” IEEE Transactions on Biomedical Engineering, vol. 57, no. 5, pp. 1158–1166, 2010. View at: Publisher Site | Google Scholar
  7. O. Ecabert, J. Peters, H. Schramm et al., “Automatic model-based segmentation of the heart in CT images,” IEEE Transactions on Medical Imaging, vol. 27, no. 9, pp. 1189–1201, 2008. View at: Publisher Site | Google Scholar
  8. J. Yang and J. S. Duncan, “3D image segmentation of deformable objects with joint shape-intensity prior models using level sets,” Medical Image Analysis, vol. 8, no. 3, pp. 285–294, 2004. View at: Publisher Site | Google Scholar
  9. X. Tao, J. L. Prince, and C. Davatzikos, “Using a statistical shape model to extract sulcal curves on the outer cortex of the human brain,” IEEE Transactions on Medical Imaging, vol. 21, no. 5, pp. 513–524, 2002. View at: Publisher Site | Google Scholar
  10. N. Duta and M. Sonka, “Segmentation and interpretation of MR brain images: an improved active shape model,” IEEE Transactions on Medical Imaging, vol. 17, no. 6, pp. 1049–1062, 1998. View at: Publisher Site | Google Scholar
  11. R. Davies, C. Twining, and C. Taylor, Statistical Models of Shape Optimisation and Evaluation, Springer, London, UK, 2008.
  12. R. C. Gonzalez and R. E. Woods, Digital Image Processing, Pearson Education, 2008.
  13. J. Rogowska, “Overview and fundamentals of medical image segmentation,” in Handbook of Medical Image Processing and Analysis, I. Bankman, Ed., Elsevier, Amsterdam, The Netherlands, 2000. View at: Google Scholar
  14. J. Canny, “A computational approach to edge detection,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 8, no. 6, pp. 679–698, 1986. View at: Google Scholar
  15. D. Marr and E. Hildreth, “Theory of edge detection,” Proceedings of the Royal Society of London—Biological Sciences, vol. 207, no. 1167, pp. 187–217, 1980. View at: Publisher Site | Google Scholar
  16. P. Kovesi, “Image feature from phase congurency,” Journal of Computer Vision Research, vol. 1, no. 3, pp. 1–26, 1999. View at: Google Scholar
  17. P. Kovesi, “Edges are not just steps,” in Proceedings of the 5th Asian Conference on Computer Vision (ACCV '02), pp. 822–827, Melbourne, Australia, 2002. View at: Google Scholar
  18. M. C. Morrone and R. A. Owens, “Feature detection from local energy,” Pattern Recognition Letters, vol. 6, no. 5, pp. 303–313, 1987. View at: Publisher Site | Google Scholar
  19. H. Gudbjartsson and S. Patz, “The Rician distribution of noisy MRI data,” Magnetic Resonance in Medicine, vol. 34, no. 6, pp. 910–914, 1995. View at: Publisher Site | Google Scholar
  20. D. Salas-González, J. M. Górriz, J. Ramírez, M. Schloegl, E. W. Lang, and A. Ortiz, “Parameterization of the distribution of white and grey matter in MRI using-stable distribution,” Computers in Biology and Medicine, vol. 43, pp. 559–567, 2013. View at: Google Scholar
  21. B. Fischl, D. H. Salat, E. Busa et al., “Whole brain segmentation: automated labeling of neuroanatomical structures in the human brain,” Neuron, vol. 33, no. 3, pp. 341–355, 2002. View at: Publisher Site | Google Scholar
  22. P. Maillard, N. Delcroix, F. Crivello et al., “An automated procedure for the assessment of white matter hyperintensities by multispectral (T1, T2, PD) MRI and an evaluation of its between-centre reproducibility based on two large community databases,” Neuroradiology, vol. 50, no. 1, pp. 31–42, 2008. View at: Publisher Site | Google Scholar
  23. A. Mayer and H. Greenspan, “An adaptive mean-shift framework for MRI brain segmentation,” IEEE Transactions on Medical Imaging, vol. 28, no. 8, pp. 1238–1250, 2009. View at: Publisher Site | Google Scholar
  24. C. R. Traynora, G. J. Barkerb, W. R. Crumb, S. C. R. Williamsb, and M. P. Richardsona, “Segmentation of the thalamus in MRI based on T1 and T2,” NeuroImage, vol. 56, no. 1, pp. 939–950, 2011. View at: Publisher Site | Google Scholar
  25. C. M. Collins, W. Liu, W. Schreiber, Q. X. Yang, and M. B. Smith, “Central brightening due to constructive interference with, without, and despite dielectric resonance,” Journal of Magnetic Resonance Imaging, vol. 21, no. 2, pp. 192–196, 2005. View at: Publisher Site | Google Scholar
  26. J. G. Sied, A. P. Zijdenbos, and A. C. Evans, “A nonparametric method for automatic correction of intensity nonuniformity in MRI data,” IEEE Transactions on Medical Imaging, vol. 17, no. 1, pp. 87–97, 1998. View at: Publisher Site | Google Scholar
  27. E. B. Lewis and N. C. Fox, “Correction of differential intensity inhomogeneity in longitudinal MR images,” NeuroImage, vol. 23, no. 1, pp. 75–83, 2004. View at: Publisher Site | Google Scholar
  28. D. W. Shattuck, S. R. Sandor-Leahy, K. A. Schaper, D. A. Rottenberg, and R. M. Leahy, “Magnetic resonance image tissue classification using a partial volume model,” NeuroImage, vol. 13, no. 5, pp. 856–876, 2001. View at: Publisher Site | Google Scholar
  29. M. S. Cohen, R. M. DuBois, and M. M. Zeneih, “Rapid and effective correction of RF inhomogeneity for high field magnetic resonance imaging,” Human Brain Mapping, vol. 10, no. 4, pp. 204–211, 2000. View at: Publisher Site | Google Scholar
  30. W. M. Wells III, W. E. L. Crimson, R. Kikinis, and F. A. Jolesz, “Adaptive segmentation of MRI data,” IEEE Transactions on Medical Imaging, vol. 15, no. 4, pp. 429–442, 1996. View at: Publisher Site | Google Scholar
  31. K. Van Leemput, F. Maes, D. Vandermeulen, and P. Suetens, “Automated model-based bias field correction of MR images of the brain,” IEEE Transactions on Medical Imaging, vol. 18, no. 10, pp. 885–896, 1999. View at: Publisher Site | Google Scholar
  32. J.-F. Mangin, “Entropy minimization for automatic correction of intensity nonuniformity,” in Proceedings of the IEEE Workshop on Mathematical Methods in Biomedical Image Analysis (MMBIA '00), pp. 162–169, Hilton Head Island, SC, USA, June 2000. View at: Publisher Site | Google Scholar
  33. A. C. Evans, D. L. Collins, S. R. Mills, E. D. Brown, R. L. Kelly, and T. M. Peters, “3D statistical neuroanatomical models from 305 MRI volumes,” in Proceedings of the IEEE Nuclear Science Symposium & Medical Imaging Conference, pp. 1813–1817, November 1993. View at: Google Scholar
  34. J. V. Hajnal, D. L. G. Hill, and D. J. Hawkes, Medical Image Registration, CRC Press, New York, NY, USA, 2001.
  35. D. Shen and C. Davatzikos, “HAMMER: hierarchical attribute matching mechanism for elastic registration,” IEEE Transactions on Medical Imaging, vol. 21, no. 11, pp. 1421–1439, 2002. View at: Publisher Site | Google Scholar
  36. E. D'Agostino, F. Maes, D. Vandermeulen, and P. Suetens, “A viscous fluid model for multimodal non-rigid image registration using mutual information,” in Proceedings of the Medical Image Computing and Computer Assisted Intervention (MICCAI '02), pp. 541–548, 2002. View at: Google Scholar
  37. J. Ashburner and K. J. Friston, “Nonlinear spatial normalization using basis functions,” Human Brain Mapping, vol. 7, no. 4, pp. 254–266, 1999. View at: Publisher Site | Google Scholar
  38. D. Rueckert, L. I. Sonoda, C. Hayes, D. L. Hill, M. O. Leach, and D. J. Hawkes, “Nonrigid registration using free-form deformations: application to breast mr images,” IEEE Transactions on Medical Imaging, vol. 18, no. 8, pp. 712–721, 1999. View at: Publisher Site | Google Scholar
  39. J. M. Fitzpatrick, D. L. G. Hill, and C. R. Maurer, “Image registration,” in Handbook of Medical Imaging: Medical Image Processing and Analysis, M. Sonka and J. M. Fitzpatrick, Eds., vol. 2, SPIE Publications, 2004. View at: Google Scholar
  40. W. R. Crum, T. Hartkens, and D. L. G. Hill, “Non-rigid image registration: theory and practice,” British Journal of Radiology, vol. 77, no. 2, pp. S140–S153, 2004. View at: Publisher Site | Google Scholar
  41. B. Zitová and J. Flusser, “Image registration methods: a survey,” Image and Vision Computing, vol. 21, no. 11, pp. 977–1000, 2003. View at: Publisher Site | Google Scholar
  42. H. Xue, L. Srinivasan, S. Jiang et al., “Automatic segmentation and reconstruction of the cortex from neonatal MRI,” NeuroImage, vol. 38, no. 3, pp. 461–477, 2007. View at: Publisher Site | Google Scholar
  43. S. M. Smith, “Fast robust automated brain extraction,” Human Brain Mapping, vol. 17, no. 3, pp. 143–155, 2002. View at: Publisher Site | Google Scholar
  44. M. Battaglini, S. M. Smith, S. Brogi, and N. de Stefano, “Enhanced brain extraction improves the accuracy of brain atrophy estimation,” NeuroImage, vol. 40, no. 2, pp. 583–589, 2008. View at: Publisher Site | Google Scholar
  45. N. R. Pal and S. K. Pal, “A review on image segmentation techniques,” Pattern Recognition, vol. 26, no. 9, pp. 1277–1294, 1993. View at: Publisher Site | Google Scholar
  46. T. F. Chan and L. A. Vese, “Active contours without edges,” IEEE Transactions on Image Processing, vol. 10, no. 2, pp. 266–277, 2001. View at: Publisher Site | Google Scholar | Zentralblatt MATH
  47. Y. Boykov and G. Funka-Lea, “Graph cuts and efficient N-D image segmentation,” International Journal of Computer Vision, vol. 70, no. 2, pp. 109–131, 2006. View at: Publisher Site | Google Scholar
  48. R. Unnikrishnan, C. Pantofaru, and M. Hebert, “Toward objective evaluation of image segmentation algorithms,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 29, no. 6, pp. 929–944, 2007. View at: Publisher Site | Google Scholar
  49. W. Cai, S. Chen, and D. Zhang, “Fast and robust fuzzy c-means clustering algorithms incorporating local information for image segmentation,” Pattern Recognition, vol. 40, no. 3, pp. 825–838, 2007. View at: Publisher Site | Google Scholar | Zentralblatt MATH
  50. M. Prastawa, An MRI segmentation framework for brains with anatomical deviations [Ph.D. thesis], University of North Carolina at Chapel Hill, 2007.
  51. F. Shi, D. Shen, P.-T. Yap et al., “CENTS: cortical enhanced neonatal tissue segmentation,” Human Brain Mapping, vol. 32, no. 3, pp. 382–396, 2011. View at: Publisher Site | Google Scholar
  52. L. Wang, F. Shi, G. Li et al., “Segmentation of neonatal brain MR images using patch-driven level sets,” NeuroImage, vol. 84, pp. 141–158, 2014. View at: Publisher Site | Google Scholar
  53. N. I. Weisenfeld and S. K. Warfield, “Automatic segmentation of newborn brain MRI,” NeuroImage, vol. 47, no. 2, pp. 564–572, 2009. View at: Publisher Site | Google Scholar
  54. J. C. Moreno, V. B. Surya Prasath, H. Proença, and K. Palaniappan, “Fast and globally convex multiphase active contours for brain MRI segmentation,” Computer Vision and Image Understanding, vol. 125, pp. 237–250, 2014. View at: Publisher Site | Google Scholar
  55. D. M. Greig, B. T. Porteous, and A. H. Seheult, “Exact maximum a posteriori estimation for binary images,” Journal of the Royal Statistical Society B, vol. 51, no. 2, pp. 271–279, 1989. View at: Google Scholar
  56. J. de Oliveira and W. Pedrycz, Advances in Fuzzy Clustering and Its Applications, John Wiley & Sons, New York, NY, USA, 2007.
  57. I. Despotovíć, E. Vansteenkiste, and W. Philips, “Spatially coherent fuzzy clustering for accurate and noise-robust image segmentation,” IEEE Signal Processing Letters, vol. 20, no. 4, pp. 295–298, 2013. View at: Publisher Site | Google Scholar
  58. D. C. Collier, S. S. C. Burnett, M. Amin et al., “Assessment of consistency in contouring of normal-tissue anatomic structures,” Journal of Applied Clinical Medical Physics, vol. 4, no. 1, pp. 17–24, 2003. View at: Google Scholar
  59. E. Vansteenkiste, Quantitative analysis of ultrasound images of the preterm brain [Ph.D. dissertation], Ghent University, 2007.
  60. M. Murgasova, Segmentation of brain MRI during early childhood [Ph.D. thesis], Imperial College London, 2008.
  61. P. A. Yushkevich, J. Piven, H. C. Hazlett et al., “User-guided 3D active contour segmentation of anatomical structures: significantly improved efficiency and reliability,” NeuroImage, vol. 31, no. 3, pp. 1116–1128, 2006. View at: Publisher Site | Google Scholar
  62. ITK-SNAP, 2009, http://www.itksnap.org/.
  63. M. Sezgin and B. Sankur, “Survey over image thresholding techniques and quantitative performance evaluation,” Journal of Electronic Imaging, vol. 13, no. 1, pp. 146–168, 2004. View at: Publisher Site | Google Scholar
  64. R. M. Haralick and L. G. Shapiro, “Image segmentation techniques,” Computer Vision, Graphics, and Image Processing, vol. 29, no. 1, pp. 100–132, 1985. View at: Publisher Site | Google Scholar
  65. N. Passat, C. Ronse, J. Baruthio, J.-P. Armspach, C. Maillot, and C. Jahn, “Region-growing segmentation of brain vessels: an atlas-based automatic approach,” Journal of Magnetic Resonance Imaging, vol. 21, no. 6, pp. 715–725, 2005. View at: Publisher Site | Google Scholar
  66. T. Weglinski and A. Fabijanska, “Brain tumor segmentation from MRI data sets using region growing approach,” in Proceedings of the 7th International Conference on Perspective Technologies and Methods in MEMS Design (MEMSTECH '11), pp. 185–188, May 2011. View at: Google Scholar
  67. M. del Fresno, M. Vénere, and A. Clausse, “A combined region growing and deformable model method for extraction of closed surfaces in 3D CT and MRI scans,” Computerized Medical Imaging and Graphics, vol. 33, no. 5, pp. 369–376, 2009. View at: Publisher Site | Google Scholar
  68. R. O. Duda, P. E. Hart, and D. G. Stork, Pattern Classification, John Wiley & Sons, New York, NY, USA, 2nd edition, 2001. View at: MathSciNet
  69. S. K. Warfield, F. A. Kaus, M. Jolesz, and R. Kikinis, “Adaptive template moderated spatially varying statistical classification,” in Proceedings of the Medical Image Computing and Computer Assisted Intervention (MICCAI '98), pp. 431–438, 1998. View at: Google Scholar
  70. C. A. Cocosco, A. P. Zijdenbos, and A. C. Evans, “A fully automatic and robust brain MRI tissue classification method,” Medical Image Analysis, vol. 7, no. 4, pp. 513–527, 2003. View at: Publisher Site | Google Scholar
  71. K. M. Pohl, J. Fisher, W. E. L. Grimson, R. Kikinis, and W. M. Wells, “A Bayesian model for joint segmentation and registration,” NeuroImage, vol. 31, no. 1, pp. 228–239, 2006. View at: Publisher Site | Google Scholar
  72. C. W. Therrien, Decision, Estimation, and Classification: An Introduction to Pattern Recognition and Related Topics, John Wiley & Sons, New York, NY, USA, 1989. View at: MathSciNet
  73. G. B. Coleman and H. C. Andrews, “Image segmentation by clustering,” Proceedings of the IEEE, vol. 67, no. 5, pp. 773–785, 1979. View at: Publisher Site | Google Scholar
  74. J. C. Dunn, “A fuzzy relative of the ISODATA process and its use in detecting compact well-separated clusters,” Journal of Cybernetics: Transactions of the American Society for Cybernetics, vol. 3, no. 3, pp. 32–57, 1973. View at: Google Scholar | MathSciNet
  75. J. C. Bezdek, Pattern Recognition with Fuzzy Objective Function Algorithms, Plenum Press, New York, NY, USA, 1981. View at: MathSciNet
  76. L. A. Zadeh, “Fuzzy sets,” Information and Control, vol. 8, no. 3, pp. 338–353, 1965. View at: Google Scholar
  77. M. N. Ahmed, S. M. Yamany, N. Mohamed, A. A. Farag, and T. Moriarty, “A modified fuzzy C-means algorithm for bias field estimation and segmentation of MRI data,” IEEE Transactions on Medical Imaging, vol. 21, no. 3, pp. 193–199, 2002. View at: Publisher Site | Google Scholar
  78. J.-H. Xue, A. Pizurica, W. Philips, E. Kerre, R. van de Walle, and I. Lemahieu, “An integrated method of adaptive enhancement for unsupervised segmentation of MRI brain images,” Pattern Recognition Letters, vol. 24, no. 15, pp. 2549–2560, 2003. View at: Publisher Site | Google Scholar
  79. S. Chen and D. Zhang, “Robust image segmentation using FCM with spatial constraints based on new kernel-induced distance measure,” IEEE Transactions on Systems, Man, and Cybernetics Part B: Cybernetics, vol. 34, no. 4, pp. 1907–1916, 2004. View at: Publisher Site | Google Scholar
  80. S. Shen, W. Sandham, M. Granat, and A. Sterr, “MRI fuzzy segmentation of brain tissue using neighborhood attraction with neural-network optimization,” IEEE Transactions on Information Technology in Biomedicine, vol. 9, no. 3, pp. 459–467, 2005. View at: Publisher Site | Google Scholar
  81. K.-S. Chuang, H.-L. Tzeng, S. Chen, J. Wu, and T.-J. Chen, “Fuzzy c-means clustering with spatial information for image segmentation,” Computerized Medical Imaging and Graphics, vol. 30, no. 1, pp. 9–15, 2006. View at: Publisher Site | Google Scholar
  82. Z. M. Wang, Y. C. Soh, Q. Song, and K. Sim, “Adaptive spatial information-theoretic clustering for image segmentation,” Pattern Recognition, vol. 42, no. 9, pp. 2029–2044, 2009. View at: Publisher Site | Google Scholar | Zentralblatt MATH
  83. I. Despotović, B. Goossens, E. Vansteenkiste, and W. Philips, “T1- and T2-weighted spatially constrained fuzzy C-means clustering for brain MRI segmentation,” in Medical Imaging: Image Processing, Proceedings of SPIE, p. 9, San Diego, Calif, USA, February 2010. View at: Publisher Site | Google Scholar
  84. B. N. Li, C. K. Chui, S. Chang, and S. H. Ong, “Integrating spatial fuzzy clustering with level set methods for automated medical image segmentation,” Computers in Biology and Medicine, vol. 41, no. 1, pp. 1–10, 2011. View at: Publisher Site | Google Scholar
  85. K. M. Pohl, W. M. Wells, A. Guimond et al., “Incorporating non-rigid registration into expectation maximization algorithm to segment MR images,” in Proceedings of the Medical Image Computing and Computer Assisted Intervention (MICCAI '02), pp. 564–572, 2002. View at: Google Scholar
  86. K. M. Pohl, Prior information for brain parcellation [Ph.D. thesis], Massachusetts Institute of Technology, 2005.
  87. E. D'Agostino, F. Maes, D. Vandermeulen, and P. Suetens, “Non-rigid atlas-to-image registration by minimization of class-conditional image entropy,” in Proceedings of the 7th International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI '04), pp. 745–753, September 2004. View at: Google Scholar
  88. M. Prastawa, J. H. Gilmore, W. Lin, and G. Gerig, “Automatic segmentation of MR images of the developing newborn brain,” Medical Image Analysis, vol. 9, no. 5, pp. 457–466, 2005. View at: Publisher Site | Google Scholar
  89. M. Kuklisova-Murgasova, P. Aljabar, L. Srinivasan et al., “A dynamic 4D probabilistic atlas of the developing brain,” NeuroImage, vol. 54, no. 4, pp. 2750–2763, 2011. View at: Publisher Site | Google Scholar
  90. C. Li, R. Huang, Z. Ding, J. C. Gatenby, D. N. Metaxas, and J. C. Gore, “A level set method for image segmentation in the presence of intensity inhomogeneities with application to MRI,” IEEE Transactions on Image Processing, vol. 20, no. 7, pp. 2007–2016, 2011. View at: Publisher Site | Google Scholar | MathSciNet
  91. D. Rivest-Hénault and M. Cheriet, “Unsupervised MRI segmentation of brain tissues using a local linear model and level set,” Magnetic Resonance Imaging, vol. 29, no. 2, pp. 243–259, 2011. View at: Publisher Site | Google Scholar
  92. M. Kass, A. Witkin, and D. Terzopoulos, “Snakes: active contour models,” International Journal of Computer Vision, vol. 1, no. 4, pp. 321–331, 1988. View at: Publisher Site | Google Scholar
  93. D. Terzopoulos, A. Witkin, and M. Kass, “Constraints on deformable models: recovering 3D shape and nonrigid motion,” Artificial Intelligence Journal, vol. 36, no. 1, pp. 91–123, 1988. View at: Publisher Site | Google Scholar
  94. L. D. Cohen and I. Cohen, “Finite-element methods for active contour models and balloons for 2-D and 3-D images,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 15, no. 11, pp. 1131–1147, 1993. View at: Publisher Site | Google Scholar
  95. S. Kichenassamy, A. Kumar, P. Olver, A. Tannenbaum, and A. Yezzy, “Gradient flows and geometric active contour models,” in Proceedings of the 5th International Conference on Computer Vision (ICCV '95), pp. 810–815, 1995. View at: Google Scholar
  96. R. Malladi, J. A. Sethian, and B. C. Vemuri, “Shape modeling with front propagation: a level set approach,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 17, no. 2, pp. 158–175, 1995. View at: Publisher Site | Google Scholar
  97. V. Caselles, R. Kimmel, and G. Sapiro, “Geodesic active contours,” International Journal of Computer Vision, vol. 22, no. 1, pp. 61–79, 1997. View at: Publisher Site | Google Scholar
  98. D. Mumford and J. Shah, “Optimal approximations by piecewise smooth functions and associated variational problems,” Communications on Pure and Applied Mathematics, vol. 42, no. 5, pp. 577–685, 1989. View at: Publisher Site | Google Scholar | MathSciNet
  99. A. Huang, R. Abugharbieh, and R. Tam, “A hybrid geometricstatistical deformable model for automated 3-D segmentation in brain MRI,” IEEE Transactions on Biomedical Engineering, vol. 56, no. 7, pp. 1838–1848, 2009. View at: Publisher Site | Google Scholar
  100. R. Kimmel, “Fast edge integration,” in Geometric Level Set Methods in Imaging, Vision, and Graphics, S. Osher and N. Paragios, Eds., Springer, New York, NY, USA, 2003. View at: Google Scholar
  101. C. Sagiv, N. A. Sochen, and Y. Y. Zeevi, “Integrated active contours for texture segmentation,” IEEE Transactions on Image Processing, vol. 15, no. 6, pp. 1633–1646, 2006. View at: Publisher Site | Google Scholar
  102. P. Mesejo, A. Valsecchi, L. Marrakchi-Kacem, S. Cagnoni, and S. Damas, “Biomedical image segmentation using geometric deformable models and metaheuristics,” Computerized Medical Imaging and Graphics, 2014. View at: Publisher Site | Google Scholar
  103. L. Wang, F. Shi, W. Lin, J. H. Gilmore, and D. Shen, “Automatic segmentation of neonatal images using convex optimization and coupled level sets,” NeuroImage, vol. 58, no. 3, pp. 805–817, 2011. View at: Publisher Site | Google Scholar
  104. L. A. Vese and T. F. Chan, “A multiphase level set framework for image segmentation using the Mumford and Shah model,” International Journal of Computer Vision, vol. 50, no. 3, pp. 271–293, 2002. View at: Publisher Site | Google Scholar | Zentralblatt MATH
  105. E. S. Brown, T. F. Chan, and X. Bresson, “Completely convex formulation of the Chan-Vese image segmentation model,” International Journal of Computer Vision, vol. 98, no. 1, pp. 103–121, 2012. View at: Publisher Site | Google Scholar | Zentralblatt MATH | MathSciNet
  106. M. S. Keegan, B. Sandberg, and T. F. Chan, “A multiphase logic framework for multichannel image segmentation,” Inverse Problems and Imaging, vol. 6, no. 1, pp. 95–110, 2012. View at: Publisher Site | Google Scholar | Zentralblatt MATH | MathSciNet
  107. S. H. Kang and R. March, “Existence and regularity of minimizers of a functional for unsupervised multiphase segmentation,” Nonlinear Analysis: Theory, Methods & Applications, vol. 76, pp. 181–201, 2013. View at: Publisher Site | Google Scholar | MathSciNet
  108. L. Wang, C. Li, Q. Sun, D. Xia, and C.-Y. Kao, “Active contours driven by local and global intensity fitting energy with application to brain MR image segmentation,” Computerized Medical Imaging and Graphics, vol. 33, no. 7, pp. 520–531, 2009. View at: Publisher Site | Google Scholar
  109. T. F. Chan, S. Esedoglu, and M. Nikolova, “Algorithms for finding global minimizers of image segmentation and denoising models,” SIAM Journal on Applied Mathematics, vol. 66, no. 5, pp. 1632–1648, 2006. View at: Publisher Site | Google Scholar | MathSciNet
  110. A. Chambolle, D. Cremers, and T. Pock, “A convex approach to minimal partitions,” SIAM Journal on Imaging Sciences, vol. 5, no. 4, pp. 1113–1158, 2012. View at: Publisher Site | Google Scholar | Zentralblatt MATH | MathSciNet
  111. E. Bae, J. Yuan, and X.-C. Tai, “Global minimization for continuous multiphase partitioning problems using a dual approach,” International Journal of Computer Vision, vol. 92, no. 1, pp. 112–129, 2011. View at: Publisher Site | Google Scholar | Zentralblatt MATH | MathSciNet
  112. T. Kapur, W. Eric, L. Grimson, W. M. Wells III, and R. Kikinis, “Segmentation of brain tissue from magnetic resonance images,” Medical Image Analysis, vol. 1, no. 2, pp. 109–127, 1996. View at: Publisher Site | Google Scholar
  113. Y. Masutani, T. Schiemann, and K. H. Hohne, “Vascular shape segmentation and structure extraction using a shape-based region-growing model,” in Medical Image Computing and Computer-Assisted Interventation—MICCAI'98: Proceedings of the 1st International Conference Cambridge, MA, USA, October 11–13, 1998, vol. 1496 of Lecture Notes in Computer Science, pp. 1242–1249, Springer, Berlin, Germany, 1998. View at: Publisher Site | Google Scholar
  114. Z. Tu, K. L. Narr, P. Dollar, I. Dinov, P. M. Thompson, and A. W. Toga, “Brain anatomical structure segmentation by hybrid discriminative/generative models,” IEEE Transactions on Medical Imaging, vol. 27, no. 4, pp. 495–508, 2008. View at: Publisher Site | Google Scholar
  115. I. Despotovic, E. Vansteenkiste, and W. Philips, “Brain volume segmentation in newborn infants using multi-modal MRI with a low inter-slice resolution,” in Proceedings of the Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC '10), pp. 5038–5041, Buenos Aires, Argentina, September 2010. View at: Publisher Site | Google Scholar
  116. J. M. Lötjönen, R. Wolz, J. R. Koikkalainen et al., “Fast and robust multi-atlas segmentation of brain magnetic resonance images,” NeuroImage, vol. 49, no. 3, pp. 2352–2365, 2010. View at: Publisher Site | Google Scholar
  117. C. Vijayakumar and D. Gharpure, “Development of image-processing software for automatic segmentation of brain tumors in MR images,” Journal of Medical Physics, vol. 36, no. 3, pp. 147–158, 2011. View at: Publisher Site | Google Scholar
  118. L. Gui, R. Lisowski, T. Faundez, P. S. Hüppi, F. Lazeyras, and M. Kocher, “Morphology-driven automatic segmentation of MR images of the neonatal brain,” Medical Image Analysis, vol. 16, no. 8, pp. 1565–1579, 2012. View at: Publisher Site | Google Scholar
  119. A. Ortiz, J. M. Gorriz, J. Ramirez, and D. Salas-Gonzalez, “Improving MR brain image segmentation using self-organising maps and entropy-gradient clustering,” Information Sciences, vol. 262, pp. 117–136, 2014. View at: Publisher Site | Google Scholar
  120. S. K. Warfield, M. Kaus, F. A. Jolesz, and R. Kikinis, “Adaptive, template moderated, spatially varying statistical classification,” Medical Image Analysis, vol. 4, no. 1, pp. 43–55, 2000. View at: Publisher Site | Google Scholar
  121. A. Ortiz, A. A. Palacio, J. M. Górriz, J. Ramírez, and D. Salas-González, “Segmentation of brain MRI using SOM-FCM-based method and 3D statistical descriptors,” Computational and Mathematical Methods in Medicine, vol. 2013, Article ID 638563, 12 pages, 2013. View at: Publisher Site | Google Scholar | MathSciNet
  122. K. Van Leemput, F. Maes, D. Vandermeulen, and P. Suetens, “A unifying framework for partial volume segmentation of brain MR images,” IEEE Transactions on Medical Imaging, vol. 22, no. 1, pp. 105–119, 2003. View at: Publisher Site | Google Scholar
  123. P. Santago and D. H. Gage, “Quantification of MR brain images by mixture density and partial volume modeling,” IEEE Transactions on Medical Imaging, vol. 12, no. 3, pp. 566–574, 1993. View at: Publisher Site | Google Scholar
  124. L. Nocera and J. C. Gee, “Robust partial volume tissue classification of cerebral MRI scans,” in Medical Imaging: Image Processing, vol. 3034 of Proceedings of SPIE, pp. 312–322, February 1997. View at: Publisher Site | Google Scholar
  125. D. L. Collins, A. P. Zijdenbos, V. Kollokian et al., “Design and construction of a realistic digital brain phantom,” IEEE Transactions on Medical Imaging, vol. 17, no. 3, pp. 463–468, 1998. View at: Publisher Site | Google Scholar
  126. IBSR, “The Internet Brain Segmentation Repository,” 2013, http://www.nitrc.org/projects/ibsr. View at: Google Scholar
  127. L. R. Dice, “Measures of the amount of ecologic association between species,” Ecology, vol. 26, no. 3, pp. 297–302, 1945. View at: Google Scholar
  128. W. R. Crum, O. Camara, and D. L. G. Hill, “Generalized overlap measures for evaluation and validation in medical image analysis,” IEEE Transactions on Medical Imaging, vol. 25, no. 11, pp. 1451–1461, 2006. View at: Publisher Site | Google Scholar

Copyright © 2015 Ivana Despotović et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

42116 Views | 7939 Downloads | 187 Citations
 PDF  Download Citation  Citation
 Download other formatsMore
 Order printed copiesOrder

We are committed to sharing findings related to COVID-19 as quickly and safely as possible. Any author submitting a COVID-19 paper should notify us at help@hindawi.com to ensure their research is fast-tracked and made available on a preprint server as soon as possible. We will be providing unlimited waivers of publication charges for accepted articles related to COVID-19.