Abstract

The analysis and processing of large data are a challenge for researchers. Several approaches have been used to model these complex data, and they are based on some mathematical theories: fuzzy, probabilistic, possibilistic, and evidence theories. In this work, we propose a new unsupervised classification approach that combines the fuzzy and possibilistic theories; our purpose is to overcome the problems of uncertain data in complex systems. We used the membership function of fuzzy c-means (FCM) to initialize the parameters of possibilistic c-means (PCM), in order to solve the problem of coinciding clusters that are generated by PCM and also overcome the weakness of FCM to noise. To validate our approach, we used several validity indexes and we compared them with other conventional classification algorithms: fuzzy c-means, possibilistic c-means, and possibilistic fuzzy c-means. The experiments were realized on different synthetics data sets and real brain MR images.

1. Introduction

Image segmentation is a very important operation in the process of treatment and analyzing images, and it is widely used in the different fields: pattern recognition, remote sensing, artificial intelligence, medical imaging, and so forth. The field of medical imaging includes several types of images: radiography (X-ray), ultrasound and magnetic resonance image [14]. These images are a very complex data, so their analysis is a challenge for researches.

In the literature, there are several methods that can segment these images. We can group them in four classes. The first one is the Thresholding; it allows to find the optimal threshold value, in order to extract the background objects in the image. In general, this approach is very sensitive to noise and ignores the spatial parameters [5, 6].

The second approach is the Contour; it allows to detect the contour of the image. This method is easy to implement, but unfortunately it is very sensitive to the noise and also to the parameters initialization, which means that it is mostly used with a pretreatment filter [710].

The third approach is the Region, which generates some methods: growing region (called ascendant) and splitting/merging (called descendants); this approach is very sensitive to the initial parameters and to the noise [1113].

The last approach is the Clustering; it is a very important operation in the process and data analysis, and it allows creating the homogeneous partitions using a similarity criterion [3, 4, 1433].

In this work, we are interested in clustering segmentation using the possibility theory combined with fuzzy theory.

The remainder of this paper is structured as follows. In Section 2, we present the clustering approach with three conventional algorithms, fuzzy c-means (FCM) (Algorithm 1), possibilistic c-means (PCM) (Algorithm 2), and possibilistic fuzzy c-means (PFCM) (Algorithm 3). In Section 3, the proposed new approach (HPFCM) is formulated to cluster the complex data set. In Section 4, we present the experimental results using different synthetics data sets and real MR images. Finally, the conclusion and perspectives are presented in Section 5.

Let / the data sets, : the number of cluster, is the matrix of membership degrees
function, is the matrix of cluster centers, : the degree of fuzzy, and : is the threshold representing
convergence error.
(1) Step .
(2)  (i) Initialize the matrix by random values in the interval and it also satisfies the condition in (3).
(3) Step .
(4) Repeat
(5)  (i) Update the matrix cluster centers in (5).
(6)  (ii) Update the matrix degree of membership in (4).
(7) Until
(8)    Stability of the matrix ,  

Let / the data sets, : the number of cluster, is the matrix of membership degrees
function, is the matrix of cluster centers, : the degree of fuzzy, : Degree of weight possibilistic,
and : is the threshold representing the convergence error.
Step .
(2)  (i) Initialize the matrix by random values in the interval .
Step .
(4)   Repeat
   (i) Update the matrix cluster centers in (8).
(6) (ii) Update the matrix degree of membership in (7).
    Until
(8) Stability of the matrix

Let / the data sets, : the number of cluster, is the matrix of membership degrees
function, is the matrix of cluster centers, : the degree of fuzzy, : Degree of weight possibilistic,
and : is the threshold representing the convergence error.
Step  .
(2)  (i) Initialize the matrix by random values in the interval .
Step  .
(4) Repeat
  (i) Update the matrix cluster centers in (12).
(6)   (ii) Update the matrix degree of membership in (10).
  Until
(8)  Stability of the matrix

2. Theories of Data Analysis

2.1. Classification Approach

Classification is a method that allows analysing and treating the complex data. It is widely used in different fields: pattern recognition, remote sensing, images processing, and artificial intelligence [3, 14, 16].

In the literature there are two types of classification: supervised and unsupervised. In general, the supervised approach uses a learning base in order to extract and produce the decision functions for classifying data. But, the unsupervised classification is used without a learning knowledge base, also called clustering [17, 22, 34, 35].

In this work, we are interested in studying clustering approach and its applications in medical image processing.

2.2. Clustering

Clustering is a process that allows dividing data into groups of similar pattern, and these groups are called clusters. In the literature, there are several theories of clustering. The first notion of fuzzy theory was proposed by Zadeh [36], and he established the basic principle of fuzzy theory using the fuzzy logic, in order to describe the uncertainty of belonging by a membership function. Then Ruspini [35] proposed the first notion of a fuzzy partition, and he considers that each cluster is a fuzzy set. Zadeh [34] proposed the conceptual framework for cluster analysis and pattern classification using the fuzzy set theory. Later, several researches were published; in order to improve the Bezdek algorithm, Rousseeuw et al. [3739] proposed the different objective functions that allow improving the effectiveness of fuzzy c-means algorithm.

To overcome the weakness of FCM to noise, Krishnapuram and Keller [23] proposed to relax the constraint of fuzzy, and they established the first possibilistic c-means algorithm (PCM). Then Barni et al. [40] showed that PCM algorithm is very sensitive to initializations and generates the coincident clusters. Timm et al. [29] proposed the possibilistic fuzzy clustering, and Pal et al. [30] proposed another possibilistic fuzzy c-means (PFCM) that can avoid the coincident cluster of PCM and be less sensitive to the noise.

2.3. Fuzzy c-Means

Fuzzy c-means algorithm was established by Bezdek [22], it allows classifying uncertain and imprecise data, and it is the most widely used in fuzzy clustering. The FCM model is the optimization problem : where is the data set, is the number of data, is the number of clusters, is the degree of fuzzy, is the degree of membership, is the center of cluster , and is a distance between and the object . Consider

Theorem 1 (FCM). If , for all , , , and data set contains at least different patterns, then and can be minimized only if

2.4. Possibilistic c-Means

Possibilistic c-means algorithm was introduced by Krishnapuram and Keller [23] to overcome the sensitivity of fuzzy c-means to noise. Their idea allows relaxing the fuzzy constraint.

The PCM is the optimization problem :

Theorem 2 (PCM). If , for all , , , , and data set contains at least different patterns, then and can be minimized only if

2.5. Possibilistic Fuzzy c-Means (PFCM)

Pal et al. [30] proposed the possibilistic fuzzy c-means algorithm (PFCM) that is based on FCM and PCM, and it generates the two membership functions concurrently; the first is a possibilistic membership () that presents the absolute degree of typicality, and the second function is the fuzzy membership () that presents the relative degree.

The PFCM is the optimization problem : where , and .

Theorem 3 (PFCM). If , for all , , , , and data set contains at least different patterns, then and can be minimized only if

Pal et al. [30] showed that PFCM can overcome the weakness of FCM to noise and also overcome the problem of coincident class of PCM.

3. The Proposed Algorithm HPFCM

3.1. HPFCM

Our new hybrid approach HPFCM is based on the previous two mathematical theories, the theory of fuzzy sets and possibility theory. Our contribution to hybrid HPFCM uses the results of the FCM algorithm as input data for the PCM algorithm. The degree of membership matrix PCM is initialized by the value of the result matrix of FCM. We present our algorithm in Algorithm 4.

Let / the data sets, : the maximum number of iterations, : the number of the clusters, : the
matrix of membership degrees function, : the matrix of cluster center, : the fuzzy degree, Fuzzy_classification(): is the
function of fuzzy classification, F_obj(): the objective function, Possibilist_classification(): the possibilistic classification
function, : Degree of weight possibilistic, and : the threshold Representing the convergence error.
(1) Step . Initialize the parameters
(2)  (i)
(3)  (ii) Initialize the matrix by center clusters
(4)  (iii) Initialize the matrix by membership degrees, random values in the interval
and it also satisfies the condition in (3).
(5) Step . Fuzzy process
(6) Repeat
(7)  
(8)  Compute: Fuzzy_classification ) // with (4) and (5)
(9)  
(10) Until
(11) Step . Possibilist process
(12)   // Initialize the matrix by
(13)   // Initialize the matrix by
(14) Repeat
(15)  
(16)  Compute: Possibilist_classification ) // with (7) and (8)
(17)  
(18) Until

4. Experimental Results

4.1. Synthetic Data Sets

Data Set . This data set consists of 10 patterns and two outliers (noise), as shown in Table 1 and as given in [30]. We applied the FCM, PCM, PFCM, and HPFCM to with initial parameters , (number of clusters), and and the center matrix with random values. The ideal (true) centroids are

Table 2 shows the results of prototypes of center clusters and using the four algorithms. For calculating the error between the results prototypes and ideal center clusters, we used the formula: , where * is the FCM, PCM, PFCM, and HPFCM. So, we can have the errors as follows: , , , and .

These error values show that our method gives the best prototypes for and . And Figure 1 shows the effectiveness of our approach.

4.2. Real Data Sets of UCI
4.2.1. Description of UCI Data Sets

In this section, we tested the performance of our method in different real data sets: Iris, Glass, Wine, and Wisconsin Breast cancer, as shown in Table 3. These data sets are referenced in Benchmark of UCI Repository of Machine Learning Data base [31].

Iris Data Set. It describes a type of Iris plant, and it contains 3 classes of 50 instances for each class. There is one class (class 1) which is linearly separable from the other, but the classes 2 and 3 are not linearly separable, and it is a four-dimensional data set [31].

Wine Data Set. The Wine is a 13-dimensional data set and contains the values of a chemical analysis of Wines grown in the same region in Italy but derived from three different cultivars [31].

Glass Data Set. Glass is a 9-dimensional data set that contains 214 objects of 6 classes representing the types of glass [31].

Breast Cancer Data Set. The Wisconsin Breast cancer is a 9-dimensional data set that contains 699 objects of 2 classes representing the types of benign and malignant cancers [31].

4.2.2. Clustering Results for UCI Data Sets

To evaluate the clustering accuracy, we used the index of Huang and Ng [32] as given in where is the number of clusters, is the number of patterns occurring in both cluster th and its true corresponding, and is the total number of objects in the data set.

In Table 4, using Iris data set, we can see that PCM algorithm gave less accuracy (0.667) than other algorithms, and the algorithms FCM and PFCM gave the values 0.893 and 0.9, respectively, and our approach gave the best accuracy (0.929). And using other data sets, Wine, Glass, and Breast, we can also see that all accuracy values of our approach are better compared to other algorithms: FCM, PCM, and PFCM.

4.3. Medical MR Images Data Sets
4.3.1. Validity Index

There are several validity indexes to evaluate the performance of a clustering method [25]. In our work, we used the sensitivity index (Se), specificity (Sp), and the classification accuracy (AC). These three indexes are based on the four parameters: true positive (TP), true negative (TN), false positive (FP), and false negative (FN), shown in Figure 2, Tables 5 and 6.(i)TP is a true positive, when a pixel labeled is correctly classified as positive.(ii)TN is a true negative, when a pixel labeled is correctly classified as negative.(iii)FP is a false positive, when a negative labeled pixel is incorrectly classified as positive.(iv)FN is a false negative, when a positive labeled pixel is incorrectly classified as negative.

Sensitivity (Se) index gives the proportion of true positives over all structures that should be segmented

Specificity (Sp) index gives the proportion of true negatives over all structures, that should not be segmented

Classification accuracy (CA) index gives the performance of a method:

4.3.2. Brainweb: Simulated Brain Database

This data set can simulate the brain MR images in three orthogonal views (transversal, sagittal, and coronal); it can also give three sequences volumes (T1, T2, and PD (proton density)), as shown in Figures 3, 4, and 5 [41], and a variety of slice thicknesses, noise levels, and levels of intensity nonuniformity [41].

Tables 7, 8, and 9 show the clustering results using the different brain tissues: gray matter (GM), white matter (WM), and cerebrospinal fluid (CSF), respectively. We used the different noise levels (0%, 1%, 3%, 5%, 7%, and 9%), in order to evaluate our approach HPFCM and we compared the accuracy results with other algorithms, FCM, PCM, and PFCM. We can see that FCM algorithm is not stable and very sensitive to noises, and PCM algorithm gave stable values, but it did not show enough performance; instead PFCM and HPFCM gave good performance and were stable. Also, we plotted the curves for visualizing the results, as can be seen in Figure 6, that our approach gave the best accuracy compared to PFCM and other conventional algorithms FCM and PCM. The clustering results for brain MR image with Ground Truth are shown in Figure 7.

4.3.3. IBSR Data Set

To evaluate our approach on real medical images, we used the IBSR data sets (Internet Brain Segmentation Repository), which is given by the Center for Morphometric Analysis of Massachusetts General Hospital [42].

Table 10 shows the clustering results of different brain tissues: GM, WM, and CSF. We did the study on 10 images from data sets IBSR; each image has 150 × 256 × 256 voxels. Also, the Ground Truth was established by the experts and it is available in IBSR [42].

According to the results of Table 11, we can see that our approach HPFCM gave the best results using different indexes: specificity (Se), sensitivity (Sp), and clustering accuracy (CA). In order to better analyze the results, we established the curve, as shown in Figure 8, so this curve shows that our approach is a better model and very efficient compared to other algorithms FCM, PCM, and PFCM. The clustering results of brain MR images of IBSR with Ground Truth are shown in Figure 9.

5. Conclusion

In this paper, we presented a new approach to clustering, which is based on the fuzzy c-means and possibilistic c-means. Our approach is used to model uncertain and imprecise data, in order to segment the brain tissue for medical images MRI. And it has been compared to three conventional clustering algorithms: FCM, PCM, and PFCM. The performance of our approach has been successfully validated on several synthetics data sets and real medical images MRI. In the future, we will integrate other theories: the genetic and evidence theory, in order to optimize our hybrid method and we can have a very robust modeling for complex data.