Mathematical Problems in Engineering

Volume 2013 (2013), Article ID 927591, 17 pages

http://dx.doi.org/10.1155/2013/927591

## A Comparative Study of Improved Artificial Bee Colony Algorithms Applied to Multilevel Image Thresholding

Department of Computer Science, Faculty of Science, Khon Kaen University, Khon Kaen 40002, Thailand

Received 19 July 2013; Accepted 3 September 2013

Academic Editor: Erik Cuevas

Copyright © 2013 Kanjana Charansiriphaisan et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

#### Abstract

Multilevel thresholding is a highly useful tool for the application of image segmentation. Otsu’s method, a common exhaustive search for finding optimal thresholds, involves a high computational cost. There has been a lot of recent research into various meta-heuristic searches in the area of optimization research. This paper analyses and discusses using a family of artificial bee colony algorithms, namely, the standard ABC, ABC/best/1, ABC/best/2, IABC/best/1, IABC/rand/1, and CABC, and some particle swarm optimization-based algorithms for searching multilevel thresholding. The strategy for an onlooker bee to select an employee bee was modified to serve our purposes. The metric measures, which are used to compare the algorithms, are the maximum number of function calls, successful rate, and successful performance. The ranking was performed by Friedman ranks. The experimental results showed that IABC/best/1 outperformed the other techniques when all of them were applied to multilevel image thresholding. Furthermore, the experiments confirmed that IABC/best/1 is a simple, general, and high performance algorithm.

#### 1. Introduction

Image thresholding is a fundamental problem when trying to extract knowledge from an image. It has been effectively applied to many types of images. The homogeneous image regions are grouped together and separated from the rest. Thresholding can be bilevel or multilevel thresholding, and both of these types can be classified into parametric or nonparametric approaches. This problem area has been receiving attention from various research groups, and hundreds of algorithms have been proposed based on that research. There are five publications in various journals that have reviewed image thresholding [1–5] from which a comprehensive review can be accessed. The literature review revealed that Otsu’s method [6] is a commonly used technique. It is a nonparametric approach and can solve both bilevel and multilevel thresholding. This method, which finds the optimal thresholds by maximizing the weighted sum of between-class variance of the classes, is called Otsu’s function [6]. However, its computation is very time consuming; the complexity of the method grows exponentially with the number of thresholds.

The research carried out by Hammouche et al. has influenced the work carried out in this paper [7]. Their research focused on solving the image thresholding problem by merging Otsu’s function with six metaheuristic techniques. The results of [7] concluded that the differential evolution (DE) [8] was the most efficient with respect to the quality of solution and because the particle swarm optimization (PSO) [9] converged more quickly. In this paper, a new family of metaheuristic techniques is used with Otsu’s function to solve the image thresholding problem. The aim of this paper is to find a single metaheuristic technique that has the same characteristics of both DE and PSO. Some researchers proposed the same criterion as [7]. Kulkarni and Venayagamoorthy [10] formulated the sensor node localization to solve the image segmentation problem using PSO and the bacteria foraging algorithm (BF) [11]. Both of these algorithms are faster than Otsu’s method for searching the optimal thresholds of multilevel image thresholding. In 2012, Ghamisi et al. [12] proposed methods for image segmentation based on two search techniques, namely, Darwinian particle swarm optimization (DPSO) [13] and the fractional-order Darwinian particle Swarm optimization (FODPSO) [14]. The results of FODPSO regarding the fitness value, standard deviation (STD), and CPU process time were better than those of its competitors, namely, DPSO, PSO, genetic algorithm (GA), and BF, especially when the segmentation level increased.

Artificial bee colony (ABC) is gaining popularity among the metaheuristic techniques because of its simplicity, generality, and remarkable performance [15–22]. It was also successfully applied to image thresholding [23, 24]. The evolution of ABC has seen continued improvements, and it has shown good performance levels in optimization problems. Balancing exploration and exploitation during a search is a successful criterion. In that regard, a family of the five modified search equations were chosen to generate the candidate solutions. The five equations were inspired by DE and GA. Those equations were embedded in the improved ABCs, that is, “ABC/best/1” and “ABC/rand/1” proposed in [25], “ABC/best/1” and “ABC/best/2” proposed in [26], and CABC proposed in [27]. Due to the fact that the equation used in ABC/best/1 in [25] is different from the equations used in ABC/best/1 in [26], this paper has renamed the algorithms in [25] to “IABC/best/1” and “IABC/rand/1,” respectively. The improved ABCs showed outstanding performance in finding solutions to the benchmarking problems. However, their ability to search the optimal image thresholding is not known. To the best of the authors’ knowledge, there is no paper published that reports this problem. The no free lunch theorem (NFLT) [28] is a major consideration in the area of optimization research. The theorem states that because the prior assumptions of the optimization problem are not known, no optimization algorithm can outperform any other algorithm. A general-purpose universal optimization algorithm has been proven to be theoretically impossible. One algorithm can outperform another if the specific problem is considered under reasonable restrictions. In this research, not only the five modified ABCs but also the standard ABC and other metaheuristic searches, that is, PSO, DPSO, and FODPSO, were implemented in the multilevel thresholding to compare their capabilities.

The remainder of the paper is organized as follows. In Section 2, the multilevel thresholding problem is formulated. Section 3 deals with the overview of ABC algorithm. Details of the various versions of the ABC algorithm are presented in Section 4. Performance evaluation and experimental results are presented in Section 5. The conclusions are given in Section 6.

#### 2. Multilevel Thresholding Problem Formulation

Searching the optimal thresholding is a method to minimize or maximize the objective function of the separating classes. Otsu’s method [6] is based on the maximization of the between-class variance. Given the image having size , where is the width and is the height, the pixels of a given picture are represented in gray levels, and these levels are in the range . The number of pixels at level is denoted by and the total number of pixels by . The gray-level histogram is normalized and regarded as a probability distribution: The total mean of the image can be defined as The multilevel thresholding with respect to the given threshold values ,, can be performed as below: where is the coordinate of a pixel, denotes the intensity level of a pixel. The pixels of a given image will be divided into classes in this regard. If is two, then is called bilevel thresholding.

The optimal threshold is the one that maximizes the between-class variance, , which can be defined by Otsu’s function: where represents a specific class in such a way that and are the probability of occurrence and the mean of class , respectively. The probabilities of occurrence of classes are defined by

The mean of each class can be given by Thus, the -level thresholding problem is transformed to an optimization problem. The process is to search for thresholds that maximize the value , which is generally defined as The computational complexity of (7) can be as large as [7]. This is an NP-hard problem, which means that the exhaustive search will take a very long time if is large.

#### 3. Standard Artificial Bee Colony Algorithm

The artificial bee colony (ABC) algorithm was inspired by the foraging behaviors of bee colonies. This algorithm was first proposed by Karaboga [29], and it is referred to as the standard ABC. The ABC contains three groups of bees: employed bees, onlooker bees, and scout bees. The number of employed bees equals the number of onlooker bees. The employed bees search for the food around the food sources from their knowledge. Then, they pass their food information to the onlooker bees. Each onlooker bee tends to select good food sources founded by the employed bees. It then searches for more food around the selected food source. The bees that abandon their food sources will become scout bees, and they must now search for new food sources. A description of the algorithm is given below.

##### 3.1. Initialization of the Population

The algorithm generates a group of food sources corresponding to the solutions in the search space. The -food sources, represented by , are produced randomly within the range of the boundaries of the -dimension variables: where and are the lower and upper bounds for the dimension , respectively. The fitness of food sources , will be evaluated. The counters that store the number of trials of each bee are set to 0 in this phase.

##### 3.2. Employed Bees Phase

In the employed bees phase, each employed bee goes to the food source in its knowledge base and finds a neighboring food source. The neighboring food source is defined by (9) as follows: where and are randomly chosen indexes: and is a uniform random number in the range .

Fitness value for a minimization problem can be assigned to the solution by (10) as follows: where is the cost value of the solution . Once is obtained, it will be evaluated and compared with . If the fitness of is equal to or better than that of , will be replaced by , and will become a new member of the population; otherwise, is retained.

##### 3.3. Calculating Probability Values Involved in Probabilistic Selection

After all of the employed bees complete their searches, the onlookers receive the information of the food sources from the employed bees. Then, they will each choose a food source to exploit depending on a probability related to the nectar amount of the food source (fitness values of the solution). That is to say, there may be more than one onlooker bee choosing the same food source if the source has a higher fitness. The probability is calculated according to (11) as follows: The lower the , the more probability that the food source is selected.

##### 3.4. Onlooker Bees Phase

After food sources have been chosen, each onlooker bee finds a new food source in its neighbourhood, following (9), just like the employed bees do. If a new food source has better food than , the new food source will replace and become a new member in the population.

##### 3.5. Scout Bees Phase

A greedy selection mechanism is employed between the old and candidate solutions of the employed and the onlooker bees phases. The trials counter of food will be reset to zero if the food source is improved; otherwise, its value will be incremented by one. In the scout bees phase, if the value of the trials counter of a food source is greater than a parameter, known as the “limit,” the food source is abandoned and the bee becomes a scout bee. A new food source will be produced randomly in the search space using (8), as in the case of the initialization phase. The trials counter of the bee will be reset to zero.

#### 4. The Various Artificial Bee Colony Algorithms

To enhance the performance of ABC, some modifications to the classical ABC algorithm have been made; the details of these modifications are as follows.

##### 4.1. The Modified Probabilistic Selection Phase for the Maximization Problem

The original ABC was designed for minimization problems. However, multilevel thresholding is a maximization problem. Thus, some modifications are required. An onlooker bee will choose a food source in a different way to (10) and (11). That is, an onlooker bee will randomly choose a food source with a probability value , which is calculated by the following form: where is obtained from Otsu’s function of the solution , is the best among the employed bees population. As per (12), a higher between-class variance computed by (4) obtains a higher probability. Thus, ABC can solve the maximization problem. That means that if the probability is calculated by (12), then the information from the elite employee bees that have higher between-class variances is passed to the onlooker bees.

##### 4.2. Various Search Equations

Search equations are usually beneficial to the performance of metaheuristic techniques. The standard ABC is good at exploration but poor at exploitation [25–27]. The modified search equations were accordingly proposed to address this problem. A family of search equations have been suggested to create a new value of the randomly selected dimension of the candidate food source . The family of search equations are as follows.

Gao et al. [26] borrowed the mutation strategies of DE [8] to modify the solution search equation and proposed two global best-guided artificial bee colony algorithms; the search equations are shown as follows: where the indices , , , and are mutually exclusive integers randomly chosen from and different for the base index , is the best individual vector with the best fitness in the current population, and is a random number in the range .

Gao and Liu [25] also included the other mutation strategies of DE [8] to ABC and proposed two strategies for improved artificial bee colony algorithms (IABC/best and IABC/rand), the equations for which are shown as follows: where the indices and are distinct integers uniformly chosen from the range , , is the best individual vector with the best fitness in the current population, and is a random number in the range .

Gao et al. [27] brought the crossover operator of GA [30] to improve ABC. Their new search equation is as follows: where the indices and are distinct integers uniformly chosen from the range , , , and is a random number in the range . This search equation is similar to the crossover operator of GA and is called CABC.

##### 4.3. Procedure of Various Artificial Bee Colony Algorithms for Multilevel Image Thresholding

This subsection shows the pseudocode of the five modified ABCs which apply to the image thresholding selection. The most important changes to these algorithms were the creation of a new candidate solution in each. Thus, the modified probabilistic selection phase and modified search equations from the previous sub-sections will be reformed in the pseudocode. In summary, IABC/best/1 is depicted as Algorithm 1. The four remaining algorithms have the same structure as of IABC/best/1, but line 09 and line 24 are different.

The Evaluate() functions at lines 02, 10, 25, and 41 are performed by Otsu’s function, (4). Line 09 and line 24 of IABC/best/1 algorithm produce new candidates using the search equation (15). The algorithm can be performed as ABC/best/1, ABC/best/2, IABC/rand/1, or CABC algorithms by replacing (15) with (13), (14), (16), or (17), respectively. When the algorithm is applied to multilevel image thresholding, is the number of thresholds. A food source represents a candidate -threshold values. Once the algorithm terminates, the best food source is the solution, and the optimal -threshold is the expected outcome. At this point in time, there are six algorithms where the standard ABC is included. Hence, the six-ABCs refers to them from now on.

#### 5. Experiments and Results

##### 5.1. Experimental Setup

The multilevel thresholding problem deals with finding optimal thresholds within the range that maximize a fitness criterion. The dimension of the optimization problem is the number of thresholds, , and the search space is . Five new metaheuristic algorithms, that is, ABC/best/1, ABC/best/2, IABC/best/1, IABC/rand/1, and CABC, have never been applied to image thresholding before. These techniques are implemented and compared with the existing metaheuristics that perform image thresholding, that is, PSO, DPSO, FODPSO, and ABC algorithms. All methods were programmed in MATLAB on a personal computer with a 3.2 GHz CPU with 8 GB RAM running on a Microsoft Windows 7 system. Experiments were conducted on 12 real images. The images, namely starfish, snow, fox, human, burro, mountain, girl, scenery, penguin, pillar, waterfall, and bridge, were taken from the Berkeley Segmentation Dataset and Benchmark [31]. Each image has a unique gray level histogram. These original test images and their histograms are depicted in Figure 1. The segmented images, with regards to their optimal thresholds (listed in Table 2) are depicted in Figure 2. A subproblem refers to a process of an image thresholding with respect to a given number of thresholds. The numbers of thresholds investigated in the experiments were two to five. Thus, the total number of sub-problems is . Each sub-problem was repeated 50 times per image, and each time is called a run.

With reference to the six-ABCs and PSO algorithms, the objective function evaluation is computed for , where is the population size and is the number of generations. A PSO particle calls Otsu’s function one time per generation. The population size in the PSO algorithm was set to 50. A bee in the six-ABCs calls Otsu’s function two times per generation their numbers of food sources were set to a half of the PSO’s size, 25. As for the DPSO and FODPSO algorithms comprising of 25 particles with 4 swarms, their objective function evaluations per generation are twice those of the six-ABCs and PSO. In this regard, the number of generations cannot be used as a metric for algorithm comparison, but the NFC is comparable. The other control parameters of PSO, DPSO, and FODPSO are tabulated in Table 1. The maximum number of generations was set to 50, 100, 150, and 150 when is 2, 3, 4, and 5, respectively.

##### 5.2. Comparing Strategies and Metrics

To minimize the effect of the stochastic nature of the algorithms on the metrics, the reported number of function calls for each sub-problem is the average or the maximum value over 50 runs. In this study, three metrics, namely, the maximum number of function calls (), successful rate (), and successful performance (), have been utilized to compare the algorithms. The convergence speed was compared by the measurement of the maximum number of function calls . A smaller means higher convergence speed. The termination criterion of an algorithm is generally to find a value smaller than the value to reach (VTR) before reaching both the maximum number of function calls and the maximum number of generations. The VTR in this research is the maximum between-class variance when the optimal thresholds were obtained from Otsu’s method. Finding the is a priority. Once all 50 runs were completed, the runs that reached the optimal solutions within the maximum number of generations were selected. These were called the successful runs. The maximum value of their NFCs represents the . The average of represents the .

The ratio for which the algorithm succeeds in reaching the VTR for each sub-problem is measured as the successful rate. is
Furthermore, the average successful rate () over the entire data set per sub-problem is calculated as follows:
where SR_{i} represents the SR of each image of each sub-problem, and represents the number of images. The highest and the lowest are the ultimate goals in an optimization process. Therefore, two individual objectives must be considered simultaneously. In order to produce a single measure, called successful performance (), the two measures mentioned above have to be combined; this gives the following:
The is the main measure in judging which algorithm performs better than others.

###### 5.2.1. Results of Otsu’s Method

The VTR and optimal thresholds provided comparative experimental results as shown in columns 3 and 4 in Table 2. The results were as same as the optimal values of the nine metaheuristic techniques. However, the number of Otsu’s function evaluations, in the case of an exhaustive search method based on Otsu’s method, is , where is the total number of gray-levels (usually 256) and is the number of thresholds. For instance, when and , the number of objective function evaluations is 256, and when and , it is 32640. Thus, the computational complexity of Otsu’s methods is which grows exponentially with the number of thresholds. The computation time of a sub-problem with took over 10 days to compute on our test machine.

###### 5.2.2. Stability Analysis

In general, the metaheuristic methods are stochastic and random searching algorithms. The results are not absolutely identical in each run of the algorithm and are influenced by the searching ability of the algorithm. As a result, it is necessary to evaluate the stability of the algorithms. The comparison of the convergence gives us valuable information in terms of the ratio representing the successful rates (SR) in Table 2.

Table 2 presents the number of thresholds, the maximum between-class variance or VTR, the optimal threshold values, successful rates () of the 50 independent runs, and the maximum number of function calls (). The average successful rate for each sub-problem is shown in Table 3. The last row and respective column of Table 3 show the averages from all of the experiments. A higher means a higher stability. From the results, the stability of each algorithm can be sorted in decreasing order by into the following order: IABC/best/1, ABC/best/2, ABC/best/1, IABC/best/2, ABC, PSO, DPSO, CABC, and FODPSO. This means that the IABC/best/1 is the most stable metaheuristic algorithm when compared to the others.

###### 5.2.3. Convergence Rate Comparison

The maximum numbers of NFC of the successful runs are listed in Table 2. A smaller means a higher convergence rate. The average of for each sub-problem is shown in Table 3. The last row and respective column of Table 3 show the average of each from all of the experiments. From the results, the convergence rate can be sorted in decreasing order by into the following order: IABC/best/1, ABC/best/1, ABC/best/2, IABC/rand/1, CABC, PSO, ABC, DPSO, and FODPSO. This means that the IABC/best/1 has the highest convergent rate when compared to the other algorithms. The results also revealed that the increases significantly as the number of thresholds increases.

###### 5.2.4. Performance Comparison

Scatter plots helped to reveal the relationships between the variables. There are four figures, one figure per threshold number; that is, Figure 3 is for , Figure 4 is for , Figure 5 is for , and Figure 6 is for . Each figure is constructed using two data sets, the successful rates (SR) versus the maximum number of function calls of all competitors from Table 2 and the average successful rates (SR_{avg}) versus the average of () from Table 3; the locations are pointed to using arrows. Those locations represent the average performances over twelve tested images of the algorithms. Some of the algorithm’s labels are not shown because its is much larger than that of the others, and therefore, they are out of range. To interpret a figure, a lower and a larger SR imply a higher performing algorithm in the search for optimal thresholds. This relationship can be seen as a point at the topleft of each scatter graph. Due to the fact that all of the scatter graph results are so similar, only information regarding the IABC/best/1 will be interpreted. In Figures 3 to 6, it was realized that the positions of IABC/best/1 are closer to the left-top corners than those of the other techniques for all the figures. This means that IABC/best/1 outperformed its competitors.

However, if there is a lot of data, the conclusion from the previous paragraph may not be so clear. Therefore, a new tool is required. The successful performance () is a measure that considers the maximum number of function calls and the successful rate simultaneously, as listed in Table 4. The best are highlighted in boldface. To rank the algorithm by successful performance, an average ranking of Friedman ranks [32] was used to compare the performance. The average ranks are shown in the last row of Table 4. A smaller rank means a better performance. We noticed that the SPs of the nine algorithms can be sorted in decreasing order by average ranking into the following order: IABC/best/1, ABC/best/1, ABC/best/2, IABC/rand/1, CABC, PSO, ABC, DPSO, and FODPSO. The best average ranking was obtained by the IABC/best/1 algorithm, which outperformed the other eight algorithms.

#### 6. Conclusions

This paper adopted the ABC algorithm to solve the maximization problem by modifying the strategy of the elite employee bees selection. Five methods for multilevel segmentation of images were presented, namely, ABC/best/1, ABC/best/2, IABC/best/1, IABC/rand/1, and CABC. These methods were used to solve the Otsu problem for examining multilevel threshold values of images and to overcome the disadvantages of previous metaheuristic methods in terms of exploration and exploitation. The performance of the algorithms was tested with 12 standard test images and compared with the ABC, PSO, DPSO, and FOPSO methods. The ranking was performed by Friedman ranks. The segmentation results obtained from various test images showed that the IABC/best/1 algorithm widely outperformed the other algorithms in terms of successful performance, computation efficiency, and stability. Another advantage of the ABC algorithm is that it has fewer control parameters than those of PSO. The experiments confirmed that IABC/best/1 is a simple, general, and high performance algorithm. Future research is to be carried out to test the feasibility of IABC/best/1 algorithm for various types of image processing applications such as remote sensing applications, automatic target recognition, and complex document analysis.

#### Conflict of Interests

The authors do not have any conflict of interests with the content of the paper.

#### Acknowledgments

This work was supported by the Higher Education Research Promotion and National Research University Project of Thailand, Office of the Higher Education Commission, through the Cluster of Research to Enhance the Quality of Basic Education.

#### References

- P. K. Sahoo, S. Soltani, and A. K. C. Wong, “A survey of thresholding techniques,”
*IEEE Transactions on Computer Vision, Graphics and Image Processing*, vol. 41, no. 2, pp. 233–260, 1988. View at Publisher · View at Google Scholar · View at Scopus - N. R. Pal and S. K. Pal, “A review on image segmentation techniques,”
*Pattern Recognition*, vol. 26, no. 9, pp. 1277–1294, 1993. View at Publisher · View at Google Scholar · View at Scopus - M. Sezgin and B. Sankur, “Survey over image thresholding techniques and quantitative performance evaluation,”
*Journal of Electronic Imaging*, vol. 13, no. 1, pp. 146–168, 2004. View at Publisher · View at Google Scholar · View at Scopus - C.-I. Chang, Y. Du, J. Wang, S.-M. Guo, and P. D. Thouin, “Survey and comparative analysis of entropy and relative entropy thresholding techniques,”
*IEE Proceedings-vision Image and Signal Processing*, vol. 153, no. 6, pp. 837–850, 2006. View at Publisher · View at Google Scholar · View at Scopus - P. Lenkiewicz, M. Pereira, M. M. Freire, and J. Fernandes, “Techniques for medical image segmentation: review of the most popular approaches,”
*Biomedical Diagnostics and Clinical Technologies*, 2011. View at Publisher · View at Google Scholar - N. Otsu, “A threshold selection method from gray-level histograms,”
*IEEE Transactions on Systems, Man and Cybernetics*, vol. 9, no. 1, pp. 62–66, 1979. View at Publisher · View at Google Scholar · View at Scopus - K. Hammouche, M. Diaf, and P. Siarry, “A comparative study of various meta-heuristic techniques applied to the multilevel thresholding problem,”
*Engineering Applications of Artificial Intelligence*, vol. 23, no. 5, pp. 676–688, 2010. View at Publisher · View at Google Scholar · View at Scopus - R. Storn and K. Price, “Differential evolution—a simple and efficient heuristic for global optimization over continuous spaces,” Tech. Rep. TR-95-012, ICSI, 1995. View at Google Scholar
- J. Kennedy and R. Eberhart, “Particle swarm optimization,” in
*Proceedings of the IEEE International Conference on Neural Networks (ICNN ’95)*, vol. 4, pp. 1942–1948, Perth, Australia, December 1995. - R. V. Kulkarni and G. K. Venayagamoorthy, “Bio-inspired algorithms for autonomous deployment and localization of sensor nodes,”
*IEEE Transactions on Systems, Man and Cybernetics C*, vol. 40, no. 6, pp. 663–675, 2010. View at Publisher · View at Google Scholar · View at Scopus - K. M. Passino, “Biomimicry of bacterial foraging for distributed optimization and control,”
*IEEE Control Systems Magazine*, vol. 22, no. 3, pp. 52–67, 2002. View at Publisher · View at Google Scholar · View at Scopus - P. Ghamisi, M. S. Couceiro, J. A. Benediktsson, and N. M. F. Ferreira, “An efficient method for segmentation of image based on fractional calculus and natural selection,”
*Expert Systems with Applications*, vol. 39, pp. 12407–12417, 2012. View at Publisher · View at Google Scholar - J. Tillett, T. M. Rao, F. Sahin, R. Rao, and S. Brockport, “Darwinian particle swarm optimization,” in
*Proceedings of the 2nd Indian International Conference on Artificial Intelligence*, pp. 1474–1487, 2005. View at Publisher · View at Google Scholar - M. S. Couceiro, R. P. Rocha, N. M. F. Ferreira, and J. A. T. Machado, “Introducing the fractional-order Darwinian PSO,”
*Signal Image and Video Processing*, vol. 6, no. 3, pp. 343–350, 2012. View at Publisher · View at Google Scholar - D. Karaboga and B. Basturk, “On the performance of Artificial Bee Colony (ABC) algorithm,”
*Applied Soft Computing Journal*, vol. 8, no. 1, pp. 687–697, 2008. View at Publisher · View at Google Scholar · View at Scopus - D. Karaboga and B. Akay, “A comparative study of Artificial Bee Colony algorithm,”
*Applied Mathematics and Computation*, vol. 214, no. 1, pp. 108–132, 2009. View at Publisher · View at Google Scholar · View at MathSciNet - N. Karaboga, “A new design method based on Artificial Bee Colony algorithm for digital IIR filters,”
*Journal of the Franklin Institute*, vol. 346, no. 4, pp. 328–348, 2009. View at Publisher · View at Google Scholar · View at MathSciNet - A. Singh, “An Artificial Bee Colony algorithm for the leaf-constrained minimum spanning tree problem,”
*Applied Soft Computing Journal*, vol. 9, no. 2, pp. 625–631, 2009. View at Publisher · View at Google Scholar · View at Scopus - D. Karaboga and C. Ozturk, “A novel clustering approach: Artificial Bee Colony (ABC) algorithm,”
*Applied Soft Computing Journal*, vol. 11, no. 1, pp. 652–657, 2011. View at Publisher · View at Google Scholar · View at Scopus - C. Zhang, D. Ouyang, and J. Ning, “An Artificial Bee Colony approach for clustering,”
*Expert Systems with Applications*, vol. 37, no. 7, pp. 4761–4767, 2010. View at Publisher · View at Google Scholar · View at Scopus - Q.-K. Pan, M. F. Tasgetiren, P. N. Suganthan, and T. J. Chua, “A discrete Artificial Bee Colony algorithm for the lot-streaming flow shop scheduling problem,”
*Information Sciences*, vol. 181, no. 12, pp. 2455–2468, 2011. View at Publisher · View at Google Scholar · View at MathSciNet - H. Shayeghi and A. Ghasemi, “Market based LFC design using Artificial Bee Colony,”
*International Journal on Technical and Physical Problems of Engineering*, vol. 3, no. 6, pp. 1–10, 2011. View at Google Scholar - M.-H. Horng, “Multilevel thresholding selection based on the Artificial Bee Colony algorithm for image segmentation,”
*Expert Systems with Applications*, vol. 38, no. 11, pp. 13785–13791, 2011. View at Publisher · View at Google Scholar · View at Scopus - E. Cuevas, F. Sención, D. Zaldivar, M. Pérez-Cisneros, and H. Sossa, “A multi-threshold segmentation approach based on Artificial Bee Colony optimization,”
*Applied Intelligence*, vol. 37, no. 3, pp. 321–336, 2011. View at Publisher · View at Google Scholar · View at Scopus - W. Gao and S. Liu, “Improved Artificial Bee Colony algorithm for global optimization,”
*Information Processing Letters*, vol. 111, no. 17, pp. 871–882, 2011. View at Publisher · View at Google Scholar · View at MathSciNet - W. Gao, S. Liu, and L. Huang, “A global best Artificial Bee Colony algorithm for global optimization,”
*Journal of Computational and Applied Mathematics*, vol. 236, no. 11, pp. 2741–2753, 2012. View at Publisher · View at Google Scholar · View at MathSciNet - W. Gao, S. Liu, and L. Huang, “A novel Artificial Bee Colony algorithm based on modified search equation and orthogonal learning,”
*IEEE Transactions on Systems, Man and Cybernetics*, vol. 99, pp. 1–14, 2013. View at Google Scholar - D. H. Wolpert and W. G. Macready, “No free lunch theorems for optimization,”
*IEEE Transactions on Evolutionary Computation*, vol. 1, no. 1, pp. 67–82, 1997. View at Publisher · View at Google Scholar · View at Scopus - D. Karaboga, “An idea based on honey bee swarm for numerical optimization,” Tech. Rep. TR06, Computer Engineering Department, Engineering Faculty, Erciyes University, 2005. View at Google Scholar
- D. E. Goldberg,
*Genetic Algorithms in Search, Optimization, and Machine Learning*, Addison-Wesley, Reading, Mass, USA, 1989. - D. Martin, C. Fowlkes, D. Tal, and J. Malik, “A database of human segmented natural images and its application to evaluating segmentation algorithms and measuring ecological statistics,” in
*Proceedings of the 8th IEEE International Conference on Computer Vision (ICCV '01)*, vol. 2, pp. 416–423, Vancouver, Canada, July 2001. View at Publisher · View at Google Scholar · View at Scopus - T. M. Beasley and B. D. Zumbo, “Comparison of aligned Friedman rank and parametric methods for testing interactions in split-plot designs,”
*Computational Statistics & Data Analysis*, vol. 42, no. 4, pp. 569–593, 2003. View at Publisher · View at Google Scholar · View at MathSciNet