- About this Journal
- Abstracting and Indexing
- Aims and Scope
- Annual Issues
- Article Processing Charges
- Articles in Press
- Author Guidelines
- Bibliographic Information
- Citations to this Journal
- Contact Information
- Editorial Board
- Editorial Workflow
- Free eTOC Alerts
- Publication Ethics
- Reviewers Acknowledgment
- Submit a Manuscript
- Subscription Information
- Table of Contents
Mathematical Problems in Engineering
Volume 2013 (2013), Article ID 927591, 17 pages
A Comparative Study of Improved Artificial Bee Colony Algorithms Applied to Multilevel Image Thresholding
Department of Computer Science, Faculty of Science, Khon Kaen University, Khon Kaen 40002, Thailand
Received 19 July 2013; Accepted 3 September 2013
Academic Editor: Erik Cuevas
Copyright © 2013 Kanjana Charansiriphaisan et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Multilevel thresholding is a highly useful tool for the application of image segmentation. Otsu’s method, a common exhaustive search for finding optimal thresholds, involves a high computational cost. There has been a lot of recent research into various meta-heuristic searches in the area of optimization research. This paper analyses and discusses using a family of artificial bee colony algorithms, namely, the standard ABC, ABC/best/1, ABC/best/2, IABC/best/1, IABC/rand/1, and CABC, and some particle swarm optimization-based algorithms for searching multilevel thresholding. The strategy for an onlooker bee to select an employee bee was modified to serve our purposes. The metric measures, which are used to compare the algorithms, are the maximum number of function calls, successful rate, and successful performance. The ranking was performed by Friedman ranks. The experimental results showed that IABC/best/1 outperformed the other techniques when all of them were applied to multilevel image thresholding. Furthermore, the experiments confirmed that IABC/best/1 is a simple, general, and high performance algorithm.
Image thresholding is a fundamental problem when trying to extract knowledge from an image. It has been effectively applied to many types of images. The homogeneous image regions are grouped together and separated from the rest. Thresholding can be bilevel or multilevel thresholding, and both of these types can be classified into parametric or nonparametric approaches. This problem area has been receiving attention from various research groups, and hundreds of algorithms have been proposed based on that research. There are five publications in various journals that have reviewed image thresholding [1–5] from which a comprehensive review can be accessed. The literature review revealed that Otsu’s method  is a commonly used technique. It is a nonparametric approach and can solve both bilevel and multilevel thresholding. This method, which finds the optimal thresholds by maximizing the weighted sum of between-class variance of the classes, is called Otsu’s function . However, its computation is very time consuming; the complexity of the method grows exponentially with the number of thresholds.
The research carried out by Hammouche et al. has influenced the work carried out in this paper . Their research focused on solving the image thresholding problem by merging Otsu’s function with six metaheuristic techniques. The results of  concluded that the differential evolution (DE)  was the most efficient with respect to the quality of solution and because the particle swarm optimization (PSO)  converged more quickly. In this paper, a new family of metaheuristic techniques is used with Otsu’s function to solve the image thresholding problem. The aim of this paper is to find a single metaheuristic technique that has the same characteristics of both DE and PSO. Some researchers proposed the same criterion as . Kulkarni and Venayagamoorthy  formulated the sensor node localization to solve the image segmentation problem using PSO and the bacteria foraging algorithm (BF) . Both of these algorithms are faster than Otsu’s method for searching the optimal thresholds of multilevel image thresholding. In 2012, Ghamisi et al.  proposed methods for image segmentation based on two search techniques, namely, Darwinian particle swarm optimization (DPSO)  and the fractional-order Darwinian particle Swarm optimization (FODPSO) . The results of FODPSO regarding the fitness value, standard deviation (STD), and CPU process time were better than those of its competitors, namely, DPSO, PSO, genetic algorithm (GA), and BF, especially when the segmentation level increased.
Artificial bee colony (ABC) is gaining popularity among the metaheuristic techniques because of its simplicity, generality, and remarkable performance [15–22]. It was also successfully applied to image thresholding [23, 24]. The evolution of ABC has seen continued improvements, and it has shown good performance levels in optimization problems. Balancing exploration and exploitation during a search is a successful criterion. In that regard, a family of the five modified search equations were chosen to generate the candidate solutions. The five equations were inspired by DE and GA. Those equations were embedded in the improved ABCs, that is, “ABC/best/1” and “ABC/rand/1” proposed in , “ABC/best/1” and “ABC/best/2” proposed in , and CABC proposed in . Due to the fact that the equation used in ABC/best/1 in  is different from the equations used in ABC/best/1 in , this paper has renamed the algorithms in  to “IABC/best/1” and “IABC/rand/1,” respectively. The improved ABCs showed outstanding performance in finding solutions to the benchmarking problems. However, their ability to search the optimal image thresholding is not known. To the best of the authors’ knowledge, there is no paper published that reports this problem. The no free lunch theorem (NFLT)  is a major consideration in the area of optimization research. The theorem states that because the prior assumptions of the optimization problem are not known, no optimization algorithm can outperform any other algorithm. A general-purpose universal optimization algorithm has been proven to be theoretically impossible. One algorithm can outperform another if the specific problem is considered under reasonable restrictions. In this research, not only the five modified ABCs but also the standard ABC and other metaheuristic searches, that is, PSO, DPSO, and FODPSO, were implemented in the multilevel thresholding to compare their capabilities.
The remainder of the paper is organized as follows. In Section 2, the multilevel thresholding problem is formulated. Section 3 deals with the overview of ABC algorithm. Details of the various versions of the ABC algorithm are presented in Section 4. Performance evaluation and experimental results are presented in Section 5. The conclusions are given in Section 6.
2. Multilevel Thresholding Problem Formulation
Searching the optimal thresholding is a method to minimize or maximize the objective function of the separating classes. Otsu’s method  is based on the maximization of the between-class variance. Given the image having size , where is the width and is the height, the pixels of a given picture are represented in gray levels, and these levels are in the range . The number of pixels at level is denoted by and the total number of pixels by . The gray-level histogram is normalized and regarded as a probability distribution: The total mean of the image can be defined as The multilevel thresholding with respect to the given threshold values ,, can be performed as below: where is the coordinate of a pixel, denotes the intensity level of a pixel. The pixels of a given image will be divided into classes in this regard. If is two, then is called bilevel thresholding.
The optimal threshold is the one that maximizes the between-class variance, , which can be defined by Otsu’s function: where represents a specific class in such a way that and are the probability of occurrence and the mean of class , respectively. The probabilities of occurrence of classes are defined by
The mean of each class can be given by Thus, the -level thresholding problem is transformed to an optimization problem. The process is to search for thresholds that maximize the value , which is generally defined as The computational complexity of (7) can be as large as . This is an NP-hard problem, which means that the exhaustive search will take a very long time if is large.
3. Standard Artificial Bee Colony Algorithm
The artificial bee colony (ABC) algorithm was inspired by the foraging behaviors of bee colonies. This algorithm was first proposed by Karaboga , and it is referred to as the standard ABC. The ABC contains three groups of bees: employed bees, onlooker bees, and scout bees. The number of employed bees equals the number of onlooker bees. The employed bees search for the food around the food sources from their knowledge. Then, they pass their food information to the onlooker bees. Each onlooker bee tends to select good food sources founded by the employed bees. It then searches for more food around the selected food source. The bees that abandon their food sources will become scout bees, and they must now search for new food sources. A description of the algorithm is given below.
3.1. Initialization of the Population
The algorithm generates a group of food sources corresponding to the solutions in the search space. The -food sources, represented by , are produced randomly within the range of the boundaries of the -dimension variables: where and are the lower and upper bounds for the dimension , respectively. The fitness of food sources , will be evaluated. The counters that store the number of trials of each bee are set to 0 in this phase.
3.2. Employed Bees Phase
In the employed bees phase, each employed bee goes to the food source in its knowledge base and finds a neighboring food source. The neighboring food source is defined by (9) as follows: where and are randomly chosen indexes: and is a uniform random number in the range .
Fitness value for a minimization problem can be assigned to the solution by (10) as follows: where is the cost value of the solution . Once is obtained, it will be evaluated and compared with . If the fitness of is equal to or better than that of , will be replaced by , and will become a new member of the population; otherwise, is retained.
3.3. Calculating Probability Values Involved in Probabilistic Selection
After all of the employed bees complete their searches, the onlookers receive the information of the food sources from the employed bees. Then, they will each choose a food source to exploit depending on a probability related to the nectar amount of the food source (fitness values of the solution). That is to say, there may be more than one onlooker bee choosing the same food source if the source has a higher fitness. The probability is calculated according to (11) as follows: The lower the , the more probability that the food source is selected.
3.4. Onlooker Bees Phase
After food sources have been chosen, each onlooker bee finds a new food source in its neighbourhood, following (9), just like the employed bees do. If a new food source has better food than , the new food source will replace and become a new member in the population.
3.5. Scout Bees Phase
A greedy selection mechanism is employed between the old and candidate solutions of the employed and the onlooker bees phases. The trials counter of food will be reset to zero if the food source is improved; otherwise, its value will be incremented by one. In the scout bees phase, if the value of the trials counter of a food source is greater than a parameter, known as the “limit,” the food source is abandoned and the bee becomes a scout bee. A new food source will be produced randomly in the search space using (8), as in the case of the initialization phase. The trials counter of the bee will be reset to zero.
4. The Various Artificial Bee Colony Algorithms
To enhance the performance of ABC, some modifications to the classical ABC algorithm have been made; the details of these modifications are as follows.
4.1. The Modified Probabilistic Selection Phase for the Maximization Problem
The original ABC was designed for minimization problems. However, multilevel thresholding is a maximization problem. Thus, some modifications are required. An onlooker bee will choose a food source in a different way to (10) and (11). That is, an onlooker bee will randomly choose a food source with a probability value , which is calculated by the following form: where is obtained from Otsu’s function of the solution , is the best among the employed bees population. As per (12), a higher between-class variance computed by (4) obtains a higher probability. Thus, ABC can solve the maximization problem. That means that if the probability is calculated by (12), then the information from the elite employee bees that have higher between-class variances is passed to the onlooker bees.
4.2. Various Search Equations
Search equations are usually beneficial to the performance of metaheuristic techniques. The standard ABC is good at exploration but poor at exploitation [25–27]. The modified search equations were accordingly proposed to address this problem. A family of search equations have been suggested to create a new value of the randomly selected dimension of the candidate food source . The family of search equations are as follows.
Gao et al.  borrowed the mutation strategies of DE  to modify the solution search equation and proposed two global best-guided artificial bee colony algorithms; the search equations are shown as follows: where the indices , , , and are mutually exclusive integers randomly chosen from and different for the base index , is the best individual vector with the best fitness in the current population, and is a random number in the range .
Gao and Liu  also included the other mutation strategies of DE  to ABC and proposed two strategies for improved artificial bee colony algorithms (IABC/best and IABC/rand), the equations for which are shown as follows: where the indices and are distinct integers uniformly chosen from the range , , is the best individual vector with the best fitness in the current population, and is a random number in the range .
Gao et al.  brought the crossover operator of GA  to improve ABC. Their new search equation is as follows: where the indices and are distinct integers uniformly chosen from the range , , , and is a random number in the range . This search equation is similar to the crossover operator of GA and is called CABC.
4.3. Procedure of Various Artificial Bee Colony Algorithms for Multilevel Image Thresholding
This subsection shows the pseudocode of the five modified ABCs which apply to the image thresholding selection. The most important changes to these algorithms were the creation of a new candidate solution in each. Thus, the modified probabilistic selection phase and modified search equations from the previous sub-sections will be reformed in the pseudocode. In summary, IABC/best/1 is depicted as Algorithm 1. The four remaining algorithms have the same structure as of IABC/best/1, but line 09 and line 24 are different.
The Evaluate() functions at lines 02, 10, 25, and 41 are performed by Otsu’s function, (4). Line 09 and line 24 of IABC/best/1 algorithm produce new candidates using the search equation (15). The algorithm can be performed as ABC/best/1, ABC/best/2, IABC/rand/1, or CABC algorithms by replacing (15) with (13), (14), (16), or (17), respectively. When the algorithm is applied to multilevel image thresholding, is the number of thresholds. A food source represents a candidate -threshold values. Once the algorithm terminates, the best food source is the solution, and the optimal -threshold is the expected outcome. At this point in time, there are six algorithms where the standard ABC is included. Hence, the six-ABCs refers to them from now on.
5. Experiments and Results
5.1. Experimental Setup
The multilevel thresholding problem deals with finding optimal thresholds within the range that maximize a fitness criterion. The dimension of the optimization problem is the number of thresholds, , and the search space is . Five new metaheuristic algorithms, that is, ABC/best/1, ABC/best/2, IABC/best/1, IABC/rand/1, and CABC, have never been applied to image thresholding before. These techniques are implemented and compared with the existing metaheuristics that perform image thresholding, that is, PSO, DPSO, FODPSO, and ABC algorithms. All methods were programmed in MATLAB on a personal computer with a 3.2 GHz CPU with 8 GB RAM running on a Microsoft Windows 7 system. Experiments were conducted on 12 real images. The images, namely starfish, snow, fox, human, burro, mountain, girl, scenery, penguin, pillar, waterfall, and bridge, were taken from the Berkeley Segmentation Dataset and Benchmark . Each image has a unique gray level histogram. These original test images and their histograms are depicted in Figure 1. The segmented images, with regards to their optimal thresholds (listed in Table 2) are depicted in Figure 2. A subproblem refers to a process of an image thresholding with respect to a given number of thresholds. The numbers of thresholds investigated in the experiments were two to five. Thus, the total number of sub-problems is . Each sub-problem was repeated 50 times per image, and each time is called a run.
With reference to the six-ABCs and PSO algorithms, the objective function evaluation is computed for , where is the population size and is the number of generations. A PSO particle calls Otsu’s function one time per generation. The population size in the PSO algorithm was set to 50. A bee in the six-ABCs calls Otsu’s function two times per generation their numbers of food sources were set to a half of the PSO’s size, 25. As for the DPSO and FODPSO algorithms comprising of 25 particles with 4 swarms, their objective function evaluations per generation are twice those of the six-ABCs and PSO. In this regard, the number of generations cannot be used as a metric for algorithm comparison, but the NFC is comparable. The other control parameters of PSO, DPSO, and FODPSO are tabulated in Table 1. The maximum number of generations was set to 50, 100, 150, and 150 when is 2, 3, 4, and 5, respectively.
5.2. Comparing Strategies and Metrics
To minimize the effect of the stochastic nature of the algorithms on the metrics, the reported number of function calls for each sub-problem is the average or the maximum value over 50 runs. In this study, three metrics, namely, the maximum number of function calls (), successful rate (), and successful performance (), have been utilized to compare the algorithms. The convergence speed was compared by the measurement of the maximum number of function calls . A smaller means higher convergence speed. The termination criterion of an algorithm is generally to find a value smaller than the value to reach (VTR) before reaching both the maximum number of function calls and the maximum number of generations. The VTR in this research is the maximum between-class variance when the optimal thresholds were obtained from Otsu’s method. Finding the is a priority. Once all 50 runs were completed, the runs that reached the optimal solutions within the maximum number of generations were selected. These were called the successful runs. The maximum value of their NFCs represents the . The average of represents the .
The ratio for which the algorithm succeeds in reaching the VTR for each sub-problem is measured as the successful rate. is Furthermore, the average successful rate () over the entire data set per sub-problem is calculated as follows: where SRi represents the SR of each image of each sub-problem, and represents the number of images. The highest and the lowest are the ultimate goals in an optimization process. Therefore, two individual objectives must be considered simultaneously. In order to produce a single measure, called successful performance (), the two measures mentioned above have to be combined; this gives the following: The is the main measure in judging which algorithm performs better than others.
5.2.1. Results of Otsu’s Method
The VTR and optimal thresholds provided comparative experimental results as shown in columns 3 and 4 in Table 2. The results were as same as the optimal values of the nine metaheuristic techniques. However, the number of Otsu’s function evaluations, in the case of an exhaustive search method based on Otsu’s method, is , where is the total number of gray-levels (usually 256) and is the number of thresholds. For instance, when and , the number of objective function evaluations is 256, and when and , it is 32640. Thus, the computational complexity of Otsu’s methods is which grows exponentially with the number of thresholds. The computation time of a sub-problem with took over 10 days to compute on our test machine.
5.2.2. Stability Analysis
In general, the metaheuristic methods are stochastic and random searching algorithms. The results are not absolutely identical in each run of the algorithm and are influenced by the searching ability of the algorithm. As a result, it is necessary to evaluate the stability of the algorithms. The comparison of the convergence gives us valuable information in terms of the ratio representing the successful rates (SR) in Table 2.
Table 2 presents the number of thresholds, the maximum between-class variance or VTR, the optimal threshold values, successful rates () of the 50 independent runs, and the maximum number of function calls (). The average successful rate for each sub-problem is shown in Table 3. The last row and respective column of Table 3 show the averages from all of the experiments. A higher means a higher stability. From the results, the stability of each algorithm can be sorted in decreasing order by into the following order: IABC/best/1, ABC/best/2, ABC/best/1, IABC/best/2, ABC, PSO, DPSO, CABC, and FODPSO. This means that the IABC/best/1 is the most stable metaheuristic algorithm when compared to the others.
5.2.3. Convergence Rate Comparison
The maximum numbers of NFC of the successful runs are listed in Table 2. A smaller means a higher convergence rate. The average of for each sub-problem is shown in Table 3. The last row and respective column of Table 3 show the average of each from all of the experiments. From the results, the convergence rate can be sorted in decreasing order by into the following order: IABC/best/1, ABC/best/1, ABC/best/2, IABC/rand/1, CABC, PSO, ABC, DPSO, and FODPSO. This means that the IABC/best/1 has the highest convergent rate when compared to the other algorithms. The results also revealed that the increases significantly as the number of thresholds increases.
5.2.4. Performance Comparison
Scatter plots helped to reveal the relationships between the variables. There are four figures, one figure per threshold number; that is, Figure 3 is for , Figure 4 is for , Figure 5 is for , and Figure 6 is for . Each figure is constructed using two data sets, the successful rates (SR) versus the maximum number of function calls of all competitors from Table 2 and the average successful rates (SRavg) versus the average of () from Table 3; the locations are pointed to using arrows. Those locations represent the average performances over twelve tested images of the algorithms. Some of the algorithm’s labels are not shown because its is much larger than that of the others, and therefore, they are out of range. To interpret a figure, a lower and a larger SR imply a higher performing algorithm in the search for optimal thresholds. This relationship can be seen as a point at the topleft of each scatter graph. Due to the fact that all of the scatter graph results are so similar, only information regarding the IABC/best/1 will be interpreted. In Figures 3 to 6, it was realized that the positions of IABC/best/1 are closer to the left-top corners than those of the other techniques for all the figures. This means that IABC/best/1 outperformed its competitors.
However, if there is a lot of data, the conclusion from the previous paragraph may not be so clear. Therefore, a new tool is required. The successful performance () is a measure that considers the maximum number of function calls and the successful rate simultaneously, as listed in Table 4. The best are highlighted in boldface. To rank the algorithm by successful performance, an average ranking of Friedman ranks  was used to compare the performance. The average ranks are shown in the last row of Table 4. A smaller rank means a better performance. We noticed that the SPs of the nine algorithms can be sorted in decreasing order by average ranking into the following order: IABC/best/1, ABC/best/1, ABC/best/2, IABC/rand/1, CABC, PSO, ABC, DPSO, and FODPSO. The best average ranking was obtained by the IABC/best/1 algorithm, which outperformed the other eight algorithms.
This paper adopted the ABC algorithm to solve the maximization problem by modifying the strategy of the elite employee bees selection. Five methods for multilevel segmentation of images were presented, namely, ABC/best/1, ABC/best/2, IABC/best/1, IABC/rand/1, and CABC. These methods were used to solve the Otsu problem for examining multilevel threshold values of images and to overcome the disadvantages of previous metaheuristic methods in terms of exploration and exploitation. The performance of the algorithms was tested with 12 standard test images and compared with the ABC, PSO, DPSO, and FOPSO methods. The ranking was performed by Friedman ranks. The segmentation results obtained from various test images showed that the IABC/best/1 algorithm widely outperformed the other algorithms in terms of successful performance, computation efficiency, and stability. Another advantage of the ABC algorithm is that it has fewer control parameters than those of PSO. The experiments confirmed that IABC/best/1 is a simple, general, and high performance algorithm. Future research is to be carried out to test the feasibility of IABC/best/1 algorithm for various types of image processing applications such as remote sensing applications, automatic target recognition, and complex document analysis.
Conflict of Interests
The authors do not have any conflict of interests with the content of the paper.
This work was supported by the Higher Education Research Promotion and National Research University Project of Thailand, Office of the Higher Education Commission, through the Cluster of Research to Enhance the Quality of Basic Education.
- P. K. Sahoo, S. Soltani, and A. K. C. Wong, “A survey of thresholding techniques,” IEEE Transactions on Computer Vision, Graphics and Image Processing, vol. 41, no. 2, pp. 233–260, 1988.
- N. R. Pal and S. K. Pal, “A review on image segmentation techniques,” Pattern Recognition, vol. 26, no. 9, pp. 1277–1294, 1993.
- M. Sezgin and B. Sankur, “Survey over image thresholding techniques and quantitative performance evaluation,” Journal of Electronic Imaging, vol. 13, no. 1, pp. 146–168, 2004.
- C.-I. Chang, Y. Du, J. Wang, S.-M. Guo, and P. D. Thouin, “Survey and comparative analysis of entropy and relative entropy thresholding techniques,” IEE Proceedings-vision Image and Signal Processing, vol. 153, no. 6, pp. 837–850, 2006.
- P. Lenkiewicz, M. Pereira, M. M. Freire, and J. Fernandes, “Techniques for medical image segmentation: review of the most popular approaches,” Biomedical Diagnostics and Clinical Technologies, 2011.
- N. Otsu, “A threshold selection method from gray-level histograms,” IEEE Transactions on Systems, Man and Cybernetics, vol. 9, no. 1, pp. 62–66, 1979.
- K. Hammouche, M. Diaf, and P. Siarry, “A comparative study of various meta-heuristic techniques applied to the multilevel thresholding problem,” Engineering Applications of Artificial Intelligence, vol. 23, no. 5, pp. 676–688, 2010.
- R. Storn and K. Price, “Differential evolution—a simple and efficient heuristic for global optimization over continuous spaces,” Tech. Rep. TR-95-012, ICSI, 1995.
- J. Kennedy and R. Eberhart, “Particle swarm optimization,” in Proceedings of the IEEE International Conference on Neural Networks (ICNN ’95), vol. 4, pp. 1942–1948, Perth, Australia, December 1995.
- R. V. Kulkarni and G. K. Venayagamoorthy, “Bio-inspired algorithms for autonomous deployment and localization of sensor nodes,” IEEE Transactions on Systems, Man and Cybernetics C, vol. 40, no. 6, pp. 663–675, 2010.
- K. M. Passino, “Biomimicry of bacterial foraging for distributed optimization and control,” IEEE Control Systems Magazine, vol. 22, no. 3, pp. 52–67, 2002.
- P. Ghamisi, M. S. Couceiro, J. A. Benediktsson, and N. M. F. Ferreira, “An efficient method for segmentation of image based on fractional calculus and natural selection,” Expert Systems with Applications, vol. 39, pp. 12407–12417, 2012.
- J. Tillett, T. M. Rao, F. Sahin, R. Rao, and S. Brockport, “Darwinian particle swarm optimization,” in Proceedings of the 2nd Indian International Conference on Artificial Intelligence, pp. 1474–1487, 2005.
- M. S. Couceiro, R. P. Rocha, N. M. F. Ferreira, and J. A. T. Machado, “Introducing the fractional-order Darwinian PSO,” Signal Image and Video Processing, vol. 6, no. 3, pp. 343–350, 2012.
- D. Karaboga and B. Basturk, “On the performance of Artificial Bee Colony (ABC) algorithm,” Applied Soft Computing Journal, vol. 8, no. 1, pp. 687–697, 2008.
- D. Karaboga and B. Akay, “A comparative study of Artificial Bee Colony algorithm,” Applied Mathematics and Computation, vol. 214, no. 1, pp. 108–132, 2009.
- N. Karaboga, “A new design method based on Artificial Bee Colony algorithm for digital IIR filters,” Journal of the Franklin Institute, vol. 346, no. 4, pp. 328–348, 2009.
- A. Singh, “An Artificial Bee Colony algorithm for the leaf-constrained minimum spanning tree problem,” Applied Soft Computing Journal, vol. 9, no. 2, pp. 625–631, 2009.
- D. Karaboga and C. Ozturk, “A novel clustering approach: Artificial Bee Colony (ABC) algorithm,” Applied Soft Computing Journal, vol. 11, no. 1, pp. 652–657, 2011.
- C. Zhang, D. Ouyang, and J. Ning, “An Artificial Bee Colony approach for clustering,” Expert Systems with Applications, vol. 37, no. 7, pp. 4761–4767, 2010.
- Q.-K. Pan, M. F. Tasgetiren, P. N. Suganthan, and T. J. Chua, “A discrete Artificial Bee Colony algorithm for the lot-streaming flow shop scheduling problem,” Information Sciences, vol. 181, no. 12, pp. 2455–2468, 2011.
- H. Shayeghi and A. Ghasemi, “Market based LFC design using Artificial Bee Colony,” International Journal on Technical and Physical Problems of Engineering, vol. 3, no. 6, pp. 1–10, 2011.
- M.-H. Horng, “Multilevel thresholding selection based on the Artificial Bee Colony algorithm for image segmentation,” Expert Systems with Applications, vol. 38, no. 11, pp. 13785–13791, 2011.
- E. Cuevas, F. Sención, D. Zaldivar, M. Pérez-Cisneros, and H. Sossa, “A multi-threshold segmentation approach based on Artificial Bee Colony optimization,” Applied Intelligence, vol. 37, no. 3, pp. 321–336, 2011.
- W. Gao and S. Liu, “Improved Artificial Bee Colony algorithm for global optimization,” Information Processing Letters, vol. 111, no. 17, pp. 871–882, 2011.
- W. Gao, S. Liu, and L. Huang, “A global best Artificial Bee Colony algorithm for global optimization,” Journal of Computational and Applied Mathematics, vol. 236, no. 11, pp. 2741–2753, 2012.
- W. Gao, S. Liu, and L. Huang, “A novel Artificial Bee Colony algorithm based on modified search equation and orthogonal learning,” IEEE Transactions on Systems, Man and Cybernetics, vol. 99, pp. 1–14, 2013.
- D. H. Wolpert and W. G. Macready, “No free lunch theorems for optimization,” IEEE Transactions on Evolutionary Computation, vol. 1, no. 1, pp. 67–82, 1997.
- D. Karaboga, “An idea based on honey bee swarm for numerical optimization,” Tech. Rep. TR06, Computer Engineering Department, Engineering Faculty, Erciyes University, 2005.
- D. E. Goldberg, Genetic Algorithms in Search, Optimization, and Machine Learning, Addison-Wesley, Reading, Mass, USA, 1989.
- D. Martin, C. Fowlkes, D. Tal, and J. Malik, “A database of human segmented natural images and its application to evaluating segmentation algorithms and measuring ecological statistics,” in Proceedings of the 8th IEEE International Conference on Computer Vision (ICCV '01), vol. 2, pp. 416–423, Vancouver, Canada, July 2001.
- T. M. Beasley and B. D. Zumbo, “Comparison of aligned Friedman rank and parametric methods for testing interactions in split-plot designs,” Computational Statistics & Data Analysis, vol. 42, no. 4, pp. 569–593, 2003.