BioMed Research International

Volume 2016 (2016), Article ID 4516376, 9 pages

http://dx.doi.org/10.1155/2016/4516376

## Segmentation of MRI Brain Images with an Improved Harmony Searching Algorithm

^{1}School of Information and Engineering, Wenzhou Medical University, Wenzhou, Zhejiang 325000, China^{2}Zhejiang ZhongLan Environment Technology Ltd., Wenzhou, Zhejiang 325000, China^{3}School of Medical Imaging, Tianjin Medical University, Wenzhou, Zhejiang 300000, China^{4}118 Hospital of the People’s Liberation Army, Wenzhou, Zhejiang 325000, China

Received 15 March 2016; Accepted 7 April 2016

Academic Editor: Yungang Xu

Copyright © 2016 Zhang Yang et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

#### Abstract

The harmony searching (HS) algorithm is a kind of optimization search algorithm currently applied in many practical problems. The HS algorithm constantly revises variables in the harmony database and the probability of different values that can be used to complete iteration convergence to achieve the optimal effect. Accordingly, this study proposed a modified algorithm to improve the efficiency of the algorithm. First, a rough set algorithm was employed to improve the convergence and accuracy of the HS algorithm. Then, the optimal value was obtained using the improved HS algorithm. The optimal value of convergence was employed as the initial value of the fuzzy clustering algorithm for segmenting magnetic resonance imaging (MRI) brain images. Experimental results showed that the improved HS algorithm attained better convergence and more accurate results than those of the original HS algorithm. In our study, the MRI image segmentation effect of the improved algorithm was superior to that of the original fuzzy clustering method.

#### 1. Introduction

With the development of image processing technology, medical imaging technology has significantly improved. A wide variety of medical images are currently being produced. Several currently available imaging approaches include computed tomography, magnetic resonance imaging (MRI), and ultrasound. These techniques are extensively used in medical diagnosis, preoperative planning, treatment, and postmanagement detection. MRI is commonly used in actual clinical diagnosis. Compared with other technologies, MRI does not employ radiation on the human body. At the same time, high-resolution imaging of human soft tissue is attained, which can be achieved in any Italian dimensional imaging [1–4]. Although MRI technology is extensively used in medicine, MRI data and images can be generated under objective and subjective reasons for data transmission as well as the environment, and other instruments produce gradation unevenness and offset field effect. Limited resolutions produce similar noise effects. Therefore, improving MRI technology is important to enhance analysis. Under MRI, the skull is relatively bright white. The range of gray values in the skull and white matter usually overlaps. The skull bone and muscle exhibit gray values similar to those of brain tissue. In a segment containing white and gray matter, the skull is resolved together with the white matter. Therefore, the accurate segmentation of MRI images is important to eliminate interference. Methods for regional enlargement is suitable for achieving clear image segmentation of the target boundaries. If the target is unclear, then the image cannot be effectively extracted. Dynamic contour models generate enhanced segmentation effects but are disadvantageous because of long computing time. Meanwhile, deformable model methods are divided into two categories, namely, the parametric deformable model and variable-level set-shape model. These methods are achieved by iterative calculation, which takes a long time. Manual determination of iteration points must first be achieved. Given the involvement of personal and subjective factors, the segmentation attained by such method is unstable. Mathematical morphological imaging of the skull for removal treatment is effective, but a suitable threshold is more difficult to achieve using such technique. Falcao et al. [5] proposed the use of a live-wire segmentation algorithm. The algorithm can provide the user effective control of the segmentation process. In the approach, the user can intervene with the results of segmentation. Another method utilizes the artificial neural network (ANN), which is composed of many processing units (nodes). The ANN can simulate the biological, particularly the massively parallel, network of the human brain learning process. Input data acquire results quickly by training under ANN theory. With such strategy, the speed of image segmentation is effectively improved. The neural network algorithm does not entail prior knowledge of the probability distribution of image gray values; consequently, segmentation results are similar to the original image [6]. The neural network method shows its unique advantages in solving a series of complex image segmentations; however, several issues arise. First, the energy function in such case falls into local minimum values for minimized images. Second, the convergence of the neural network is related to the data; thus, a suitable value for testing network inputs is needed. The fuzzy clustering method is more extensively applied for image segmentation. The fuzzy clustering algorithm has the following advantages: the algorithm avoids the issue of threshold setting and does not entail human manipulation. Furthermore, the fuzzy clustering algorithm is particularly suitable for fuzzy and uncertain images. In this study, the fuzzy -means clustering algorithm was selected to segment MRI brain images. Scholars discovered that the effect of the initial value of the cluster centers of the fuzzy clustering algorithm is relatively large. The characteristics of a nonconvex function involve several local minima; thus, the initial value of FCM will fall into the local minima. This study utilized the rough set to compute for the initial value of the FCM.

The harmony searching (HS) algorithm was developed by Korean scholars. Geem et al. [7, 8] proposed a kind of intelligent optimization algorithm in 2001. The algorithm describes the process of musical improvisation; different musical tones are applied to a harmony vector to search for a harmony randomly. Then, the process attains an optimal harmony. Jang et al. [9] used the Nelder-Mead simplex HS algorithm, whereas Mahdavi et al. [10] adopted the adaptive HS algorithm (IHS). Omran and Mahdavi compared the performances, parameters, and noise effects related to the original HS algorithm [11], IHS, and global-best HS algorithm. H.-Q. Li and L. Li [12] employed the genetic algorithm and the HS algorithm to explore the three functions of Rastrigin, Griewank, and Sphere. Liang and coworkers [13, 14] adjusted certain parameters to improve the HS algorithm and used a hybrid GA-HS algorithm to solve the critical sliding slope problem. Cheng et al. [15] adopted the HS algorithm with several other heuristic optimization algorithms for earth slope stability analysis. Dong et al. [16] proposed the HS -means clustering algorithm to change WEB text categorization. Bezdek [17] utilized an adaptive adjustment of parameters on the improved HS algorithm to solve the anomaly detection problems in digital images of biological tissue.

In recent years, the HS algorithm has been adopted in several applications. However, the algorithm exhibits several disadvantages. The HS algorithm operates with weak robustness, considerable randomness, lack of specific direction, and slow convergence speed; it easily falls into the local optimal solution. The problem can be attributed to the search mechanism of the HS algorithm. This study proposed an improved HS algorithm for MRI brain image segmentation to overcome the aforementioned disadvantages. We used the rough set and memory bank of the HS algorithm together with the concept of rough set upper and lower boundary correction HS algorithm of the “optimal” and “worst” harmonies. By doing so, we prevented the HS algorithm convergences from falling into the local optimum. The HS algorithm should be employed to obtain the number of optimal solutions as the initial value for the average fuzzy clustering algorithm. This strategy would overcome the random determination of the initial value of the fuzzy clustering algorithm. Experimental results showed that the proposed algorithm achieved perfect convergence, and the segmentation effect was ideal for the MRI brain images. Besides ANN and fuzzy clustering, ensemble learning [18, 19], feature ranking [20], and samples selection [21] were also employed in the biomedical research.

#### 2. Harmony Searching Algorithm

The HS algorithm has been proposed as a new algorithm for the study of musical play. Each musician produces individual tones that can generate vector values. If the music produced is pleasant sounding, then the tone is recorded and tools are employed to generate a better harmony in the subsequent attempt. Musical harmony is analogous to the optimal solution vector, whereas the player riffs correspond to the optimization techniques in the local and global search programs. The HS algorithm uses a random search that selects probabilities and adjusts the pitch without information derived from the harmony. Compared with early heuristic optimization algorithms, the HS algorithm is conceptually straightforward, utilizing less mathematical expressions and a few parameters for the random search of theoretical values. Moreover, the algorithm can be more easily optimized for various engineering problems. These theoretical ideas can be adopted to formulate the solution vector , which refers to the evaluation function for . The HS algorithm is mainly divided into the following steps.

*Step 1 (initialization parameter). *The HS algorithm includes a series of important parameters, such as the number of iterations ; the harmony memory data base ; the harmony memory probability values and ; the fine-tuning probability BW; the harmony memory size ; the dimension parameter optimization problem ; and the upper and lower boundaries and , respectively.

*Step 2 (harmony memory initialization). *A harmony database is adopted to store the HMS of random harmonic vectors. The random harmonic vectors by weight of each dimension on the upper and lower boundaries and , respectively, are expressed as follows:The HM matrix expression is as follows:

*Step 3 (new harmony generation). *In accordance with the change in objective function value, the adaptive setting of the harmony memory considers the probability HMCR. By the maximum and minimum sounds of the initial setup, probability dynamic adjustment PAR is achieved. After parameter adjustment, learning by the differences in operation, pitch adjustment, and random mutation process, new harmonic solution vectors are created.

The process involved is as follows: *For *%* Beginning,* *if * HMCR* then* %* memory consideration* *if ** then* %* pitch adjustment* *else* %* random selection* *end if*

*Step 4 (memory bank updating). *The new harmony is regarded as the worst harmony in the database, and the most optimal update for the worst harmony is utilized in the database.

*Step 5 (termination conditions). *The current number of iterations is determined to achieve the maximum number of iterations. Once the number of maximum iterations is achieved, the terminate iteration cycle is commenced through Steps 3 to 5.

#### 3. Harmony Searching Algorithm Improvement

Compared with other optimization algorithms, the HS algorithm is superior on the basis of the following reasons. (1) The HS algorithm requires minimal mathematical criteria and does not entail variable initialization. (2) The entire search process of the HS algorithm assumes a completely random pattern without considerable manual intervention. (3) The HS algorithm considers the entire available acoustic memory information to create a new harmony vector. Given these advantages, the HS algorithm has been the focus of attention of many foreign scholars since 2001.

The randomness of the algorithm results in low precision. The HS algorithm is mainly adopted to improve the accuracy of the optimization problem. The HS algorithm is a strong randomness heuristic algorithm, has a simple structure, is easily operated, and involves only a few parameters and other characteristics. However, the HS algorithm also adopts sensitive parameters and generates slow convergence defects, thereby entailing further research on enhancements. The HS algorithm uses a few paramount parameters that directly affect the algorithm. These parameters include the harmony memory probability values and and the fine-tuning probability BW. In this study, the accuracy of the HS algorithms is improved to prevent attaining a premature local optimum. In this regard, the following enhancements were applied.

##### 3.1. Construction of a New Harmony HM Database

In the original harmony algorithm, harmony memory acquisition is random, thereby entailing a relatively large stochastic algorithm. This effect reduces the accuracy of the algorithm. In this study, a rough set was employed on the upper and lower boundaries to establish a new harmony HM database. Rough set theory was adopted to reduce the randomness of the harmony memory database and improve the latter’s accuracy.

*Step 1. *The relationship was applied, where is the clustering center, to establish the initial average .

*Step 2. *With the data points , , the limits of the upper and lower boundaries, and , respectively, were almost reached. and are the limits of the clustering center . was adopted to denote the distance between two points .

*Step 3. *If corresponds to the extreme minimum value, then must be close to . If is less than a given threshold, then ; otherwise, .

*Step 4. *A lower limit is established on the matrix, as follows:According to the previously presented steps, alternate data were used for preliminary screening to establish the harmony algorithm. However, more accurate data were required by the algorithm. Thus, the -nearest neighbor (KNN) algorithm was employed to attain the appropriate harmony memory matrix. denotes a given clustering center based on prior knowledge.

The KNN method [22] was originally proposed in 1968 by Cover and Hart. The KNN is a theoretically more mature classification algorithm. The core idea of the KNN algorithm is simple: if a sample feature vector space most similar (the nearest feature vector space) to the sample belongs to the major category, then the sample likely belongs to such category. The KNN method for the decision-making category is based solely on the nearest category or several categories of samples to which samples are designated to. The traditional KNN algorithm has been referred to as an example-based learning classification algorithm. By comparing each training sample, users find the text to be classified with the most similar text. Finally, the text that contains the greatest number of similar categories is selected and classified as category text. The related mathematical expression is as follows:where is the feature vector, corresponds to similarity, and denotes the classification properties. If belongs to , then the value of the function is 1; otherwise, the value is 0. Herein, we used this kind of thinking process for classification.

In distributing the matrix sample items across the class space, we applied Euclidean distance as a distribution rule as follows:

If , then each area involves a clustering center. The aforementioned methods were adopted to establish a suitable search memory database HM as follows:

*Step 5. *When all the optional data maximum and minimum values were less than the threshold , the loop was terminated. Otherwise, Steps 2 to 4 were repeated to establish the appropriate database matrix of harmony. Herein, we considered .

Matrix was established by using the new harmony matrix of rough set theory. By using rough set theory to establish a harmony matrix principle instead of a random matrix, we avoided the poor robustness and randomness of the HS algorithm.

##### 3.2. Probability PAR Adjustment

A study on the HS algorithm revealed that probability PAR tuning and volume BW are set randomly or by experience. In such case, no change in the convergence process is achieved. In fact, the effect of these two parameters on the convergence of the algorithm is relatively large, particularly in the latter part of the run. The original HS algorithm is not concerned with this aspect; it is not conducive for a fast algorithm that converges to the global optimum. In this study, the PAR and BW parameters in the original HS algorithm were improved to avoid falling into the local optimum.

In the HS algorithm, adjusting the probability PAR is also an important component. In the literature [10], the value of a small PAR has been shown to enhance the local search ability of the algorithm. By contrast, the value of a larger PAR is beneficial for adjusting the search area. The expression is shown as follows:where is the iteration number and is the current number of iterations. In this study, the global search algorithm was improved by introducing a feedback mechanism and moving a step length. The number of iterations was also updated. To update the probability of the harmony memory database and step length, we adopted the following expression:The times moving steps were expressed as follows:

By improving the HS algorithm using dynamic tone control, adjustable probability PAR values and bandwidth BW were attained, overcoming the shortcomings in probability PAR value and bandwidth generated by the fixed tone control in the basic HS algorithm. Compared with other algorithms, whether on test function or vector search solutions, the enhanced HS algorithm exhibited a better performance.

##### 3.3. Termination Conditions

At the maximum or minimum harmony database values less than the threshold , the loop was terminated. Otherwise, the original HS algorithm was repeated from Steps 2 to 4.

#### 4. Fuzzy Clustering Segmentation

The FCM clustering algorithm was proposed using fuzzy set theory. The FCM uses fuzzy set theory for classification. Data under a certain degree of categorization is divided into various types, and cluster centers are calculated in accordance with all the updated data objects of each category. This ambiguity makes the classification process of the FCM algorithm better reflect the actual data distribution, particularly for the treatment of overlap between all categories.

FCM clustering image segmentation treats pixels in an image as a cluster sample and the entire diagram as a sample set; each pixel feature vector is extracted from the image and regarded as the sample; then the pixels in the feature space are clustered. In essence, the pixels with similar characteristics are grouped in an aggregate class, whereas the pixels with dissimilar features are distributed into different classes. Finally, each pixel is completely tagged to image segmentation.

In the fuzzy means clustering algorithm (FCM), the initial value setting is a more important direct effect of segmentation speed, accuracy, and effectiveness. Before starting, the cluster number must be given first. However, in the absence of human intervention and prior knowledge of the image, such as in an automated system, determining the cluster number is a difficult task. Therefore, values based on image segmentation problems are difficult to determine under fuzzy clustering. In traditional FCM, the initial value is random. Thus, the randomness of the algorithm is high and a local optimal solution is attained.

In this study, the initial value is regarded as the number of optimal solutions obtained by the HS algorithm; the algorithm can achieve a favorable result by avoiding the local optimal solution. The MRI brain image segmentation effect attained by the improved algorithm is better than that achieved through the traditional FCM. A previous study [17] promoted the objective function of the FCM clustering algorithm; the related expression is as follows:where , is the fuzzy index, and is the distance between the clustering center and clustering objects. In this study, the Euclidean metric distance was adopted to compute the gray difference between any point and the cluster center. The Euclidean metric distance can be calculated with minimum steps.

The minimum value refers to the direction of clustering in under the condition of the constraint . By using the Lagrangian approximation solution, the degree of membership and cluster center under the extreme value are calculated as follows:

Equations (9) and (11) were utilized by continuous iterative optimized clustering. Each iteration was adopted to calculate the membership degree matrix and cluster center until convergence was reached.

The detailed steps for FCM calculation are as follows:(1)The optimal value of was obtained using the improved HS algorithm and rough set theory as the initial value for the FCM algorithm.(2)The clustering center vector was initialized.(3)On the basis of (10), the membership degree matrix was updated, where denotes the iteration number.(4)Equation (11) was employed to update the clustering center .(5)The number of iterations or error parameter when or was determined. Then, Steps 3 to 5 were repeated until the loop was terminated.

#### 5. Experiment

In a simulation experiment, the improved harmonic search algorithm and the original harmony algorithm were used to obtain the optimal, worst, and average values for MRI brain images 1–4 (MRI1–4).

In the data index, all values of the improved HS algorithm were superior to those of the original HS algorithm. The improved HS algorithm also obtained a better optimal solution. The different values were computed using Euclidean distance. The quantitative units were expressed to 10^{5}.

In the experimental data (Table 1), a smaller optimal value indicates a nearer distance to the clustering center and a more accurate selection of the cluster center. As the average value approaches the optimal value, the more optimal condition of the cluster center is achieved. More precise cluster centers attain better segmentation effects.