Biologically Inspired Methods for Imaging, Cognition, Vision, and Intelligence
View this Special IssueResearch Article  Open Access
Wei Mao, Hengyou Lan, Haoru Li, "A New Modified Artificial Bee Colony Algorithm with Exponential Function Adaptive Steps", Computational Intelligence and Neuroscience, vol. 2016, Article ID 9820294, 13 pages, 2016. https://doi.org/10.1155/2016/9820294
A New Modified Artificial Bee Colony Algorithm with Exponential Function Adaptive Steps
Abstract
As one of the most recent popular swarm intelligence techniques, artificial bee colony algorithm is poor at exploitation and has some defects such as slow search speed, poor population diversity, the stagnation in the working process, and being trapped into the local optimal solution. The purpose of this paper is to develop a new modified artificial bee colony algorithm in view of the initial population structure, subpopulation groups, step updating, and population elimination. Further, depending on oppositionbased learning theory and the new modified algorithms, an improved type grouping method is proposed and the original way of roulette wheel selection is substituted through sensitivitypheromone way. Then, an adaptive step with exponential functions is designed for replacing the original random step. Finally, based on the new test function versions CEC13, six benchmark functions with the dimensions and are chosen and applied in the experiments for analyzing and comparing the iteration speed and accuracy of the new modified algorithms. The experimental results show that the new modified algorithm has faster and more stable searching and can quickly increase poor population diversity and bring out the global optimal solutions.
1. Introduction
It is well known that algorithms for solving various characteristics optimization problems can be classified into different groups, such as populationbased algorithms, stochastic algorithms, deterministic algorithms, and iterative algorithms. An algorithm is called populationbased [1] if one works with a group of solutions and tries to improve them. Two important classes of populationbased optimization algorithms are exactly evolutionary algorithms and swarm intelligencebased algorithms [2]. Swarm intelligence is an innovative artificial intelligence technique with collective behavior of selforganized systems [3]. Since many swarm intelligence algorithms, such as genetic algorithm (GA) [4], particle swarm optimization (PSO) [5], ant colony optimization (ACO) [6], and biogeographybased optimization [7], have simplicity, ease of implementation, outstanding performance, and other advantages [8], they have shown great success in solving some nonconvex, discontinuous, or nondifferentiable optimization problems. However, these intelligence algorithms are sensitive to value and precision. Thus, inspired by the behavior of honey bees, Karaboga [9] introduced basic artificial bee colony algorithm (ABC) in 2005 and constituted one of the most prominent approaches in the field of beeinspired algorithms. Further, in consideration of the solution reaching speed, the success rate, and the performance rate, ElAbd [10] provided a complete performance assessment of ABC and compared it with the widely known differential evolution (DE), GA, heuristic algorithms, PSO, and other foraging algorithms (e.g., bacterial algorithm, ACO, and bacterial foraging optimization) by using the wellknown benchmark functions in [11].
It was claimed that ABC is the most successful algorithm for multimodal and hybrid functions [12]. This is because ABC has no demand to the objective function, constraint, and external information and is only based on fitness function in the search process [13]. Further, ABC has the following advantages. (i) The mechanism of multiple roles: using different methods, bees adjust quality of the solutions spontaneously, so as to adapt to the next search process [14]. (ii) The cooperative working mechanism: according to the information from other bees, bees decide whether to find the optimal solution with larger probability [15]. (iii) The strong robustness: the search rules are not certain but are probabilistic and have excellent robustness and a wide range of applicability [16]. (iv) The stability: even if the individual fails, the entire swarm can still complete the task [17]. (v) Less control parameters, simple operation, and ease of implementation [18]: indeed, ABC has been shown to be very competitive with respect to other stateoftheart foraging and evolutionary algorithms.
However, there exists still an insufficiency to ABC because ABC does well in exploration but is poor at exploitation. As for the improvement and development of ABC, Karaboga and Gorkemli [19] proposed a new update rule for onlooker bees in the hive to improve the local search and convergence characteristics of ABC. Inspired by PSO, Imanian et al. [20] changed the update rule of basic ABC to increase the convergence speed for solving high dimensional and continuous optimization problems. Wang et al. [21] proposed multistrategy ensemble artificial bee colony algorithm (MEABC) to improve the local and global search capability of basic ABC and tested the performance of MEABC by using basic, shifted, and rotated benchmark functions. Gao et al. [22] developed new search equations to adjust exploration and exploitation capability of ABC. In a different approach for ABC, Das et al. [23] proposed a learning routine based on fitness and proximity stimuli and tested the method with standard benchmark functions. Zang et al. [24] designed a logarithmic function adaptive step instead of the original random step.
Moreover, in dealing with some complex problems by applying ABC, there are some defects such as slow search speed, poor population diversity, stagnation in the working process, and trapping into the local optimal solution [13]. Recently, ABC has been extended and improved by many researches. But since ABC is relatively new, the researches in the literatures lack systematicness and are scattered. See, for example, [1–3, 12–30] and the references therein. In 2012, Li et al. [31] pointed out that “ABC has no mechanism to use the global information in the search space, so it easily results in a waste of computing power and gets trapped in local optima”. Further, the authors proposed a novel algorithm (named as DEABC, i.e., differential evolution artificial bee colony algorithm), which synthesizes DE and ABC and enhances individuals by sharing information between DE population and bee colony. For related works, one can see [1, 30] and the references therein.
Motivated and inspired by the above works, a new modified artificial bee colony (MABC) algorithm shall be constructed based on adaptive step with exponential functions. By using oppositionbased learning theory and an improved type grouping method, the initial population for MABC will be given, and the original way of roulette wheel selection shall be substituted by sensitivitypheromone way. Specifically, an adaptive step with exponential functions will be designed to replace the original random step. In order to verify the validity of the improved algorithm, MABC, it is compared with DEABC [30], novel artificial bee colony algorithm (NABC) [8], and ABC; the experiment results tested with six wellknown benchmark optimization functions, which are chosen from new test function versions CEC13, show that MABC is superior to ABC, NABC, and DEABC.
The rest of this paper is organized as follows. Section 2 gives a brief introduction to ABC. The new MABC is presented and analyzed in Section 3. Section 4 presents and discusses the experimental results of six benchmark functions with the dimensions and , respectively. Finally, the conclusion is drawn in Section 5.
2. Brief Review to ABC
In this section, a brief review on ABC is going to be given.
2.1. Thoughts of the Algorithm
A honey bee colony can successfully discover the highest quality food sources in nature. Hence, the idea of ABC comes from intelligent foraging behavior of honey bees to finding good solutions for solving optimization problems. In a general way, according to the ways of searching food, the colony of bees is divided into three kinds: employed bees, onlooker bees, and scout bees. The employed bees are responsible for exploiting the nectar sources. They explore the beforehand food source position and give the quality information of the food to the onlooker bees. The onlooker bees wait in the hive and decide to exploit a food source based on the information shared by the employed bees. In order to find a new nectar source, the scout bees randomly search environment either depending on an internal motivation or based on possible external clues [32]. The position of a nectar source implies a possible solution of the optimization problems, and the profitability of a nectar source corresponds to the quality (fitness) of the possible solution. Each nectar source is exploited by only one employed bee. In other words, the number of nectar sources equals the number of employed bees or onlooker bees [25]. In this process, the employed bees maintain good solution, the onlooker bees improve convergence speed, and the scout bees enhance the ability to remove local optimum [26, 31].
2.2. ABC Iteration Steps
From [24, 33], it follows that main iteration steps of ABC can be listed as follows.
Step 1 (initialization). Randomly generate solutions (i.e., food sources) as initial population in a dimension searching space, where is the number of food sources, which equals the half of the colony size, is a dimensional solution vector and the th food source in the initial population for , and also denotes the number of optimization parameters.
Step 2 (renewing population). In the stage of collecting honey, each employed bee produces a new nectar source within the neighborhood of the food source. After comparing the new nectar source with the old ones, the high probability will be memorized. Next, in the stage of follow, every onlooker bee evaluates the profitability of nectar sources taken from all employed bees and then chooses a food source at a certain probability. As in the case of the employed bees, she produces a modification on the source position in her memory and keeps the better nectar source. The regeneration of nectar sources in these two stages is based on the following formula:where , , , and is a random number, which controls the generation range of neighborhood of . With the search being close to optimal solution, the range of neighborhood will become smaller and smaller.
Step 3 (nectar source selection). In the stage of follow, the onlooker bees choose food source through comparing the probability which is computed by the fitness value. The nectar sources of high probability are selected in large probability. And the probability of being selected for the food sources is calculated as follows:where is the fitness (profitability) value of the th solution, which is obtained by the following equation:where is the objective function value for , which is specific for the optimization problem. If the new food source position has a quality equal to or better than the old one, then the old food source position is replaced by the new one. Otherwise, the old one is retained, which is the same as the employed bees stage.
Step 4 (population elimination). If a solution has not been improved significantly through a predetermined number of trials, called “max iteration,” then the solution is regarded as falling into a local optimal solution and their original position will be abandoned. Thus, the corresponding employed bees will become scout bees and a new solution instead of the eliminated solution is randomly generated, which can be expressed as follows:where , are the same as in (1) and and denote the th individual maximum and the th individual minimum values, respectively.
Based on the above iteration steps, the process of ABC [12] can be shown in Algorithm 1.

3. New MABC
From the above introductions, it follows that ABC has many advantages, such as simplicity, ease of implementation, and outstanding performance. However, the algorithm steps indicate that ABC has the following obvious flaws:(i)The randomly generated initial population will lead the solution to be random distribution in the solution space. Hence, the searching ability of ABC will be affected directly.(ii)All individuals begin to search directly in the whole solution space, which will reduce searching efficiency.(iii)The population is regenerated by random step length. When the algorithm performs in the neighborhood of optimal solutions, the searching range will be severely restricted. Thus, the convergence rate and optimization precision of ABC will be influenced.
In order to overcome the existing deficient problem of ABC and related algorithms, the purpose of this paper is to introduce and study a new MABC and to improve the initial population structure, the population grouping, and the population regeneration of ABC.
3.1. OppositionBased Learning of Population Initialization
ABC is relatively sensitive to the initial population construction. Therefore, the search range of ABC will be restricted and the global search ability of ABC will also be influenced when the initial population distribution in the search space or the local area is random. However, the initial population constructed by using oppositionbased learning method has the diversity as much as possible, is better representative, and meets the accuracy of experimental requirements (see, e.g., [6, 13, 34, 35] and the references therein).
Based on the oppositionbased learning method, the initial population for MABC will be constructed in this paper. Firstly, an initial solution is given, and next the corresponding reverse initial solution to the initial solution is produced. Then, the two kinds of solutions are sorted. Finally, the better fitness solution is chosen as the initial population. This method will help to enhance the efficiency of MABC and to improve the quality of solutions [34]. The processes can be listed as shown in Algorithm 2.

3.2. Type Subpopulation Grouping
Intelligent optimization algorithms commonly use general grouping method as follows. Firstly, the population fitness function value of each individual is calculated. Secondly, the fitness function value is arranged in descending order. Then, the entire ordered population is divided into subpopulations, and the process of the grouping [24] is as shown in Table 1. However, this grouping method still causes some problems: the relative advantage of individuals is divided into the same species and their disadvantage is concentrated in other uncertain groups. This phenomenon will make very large differences between subpopulations, and this grouping method is not conducive to the expansion of population diversity.

To reduce the difference between each group (see [24]), a new grouping method is presented, which is called type subpopulation grouping. Similarly, the fitness function value is arranged in descending order. But, relative to the general grouping method, the order of the value in the new groups is different, and the process of new grouping is listed in Table 2.

Indeed, first of all, the fitness function values of each individual of population are calculated, and the fitness function values in descending order are ranked. Next, the entire population is divided into subpopulations . Let be the th individual for . If the solutions due to type subpopulation grouping are odd rounds and there are a large number subpopulations, then there will be more differences between each group. Thus, the final round can be abandoned through type subpopulation grouping. Now one can see that type subpopulation grouping can overcome the disequilibrium between groups, making the subpopulation have a little difference, which is good to the expansion of population diversity.
Furthermore, a new population will be recomposed by all subpopulations when each subpopulation group is less than max iteration of intragroup. And then the fitness function values of each population are recalculated and are arranged according to the descending order. Only one of the individuals of equivalent fitness function values is kept. Let the number of remaining individuals at last be . Then the population can be regenerated as follows:
3.3. SensitivityPheromone Option
Since the onlooker bees chosen by roulette way are too greedy and the population diversity is deduced too quickly, it is easily plunged into local optimum by using ABC for solving function optimization problems. In applications, the roulette way is practically affected by the population diversity.
Thus, it is an important idea for sensitivity to replace the roulette way. Sensitivity is put forward by free search algorithm [35] with pheromone (related to optimization of adaptive value) to selected area, which can be expressed as follows:(i)Calculate the values of fitness to the individual .(ii)For each , calculate the th pheromone where stands for the maximum values of for .(iii)Generate the th individual sensitivity randomly: .(iv)Find out sensitivity area matching individual of the first and randomly find to meet .
By the above model, it is easy to see the following: (i) since the sensitivitypheromone is random, any area can be searched in theory, and the local optimum can largely be avoided by using the model. (ii) The fourth step of the model is to make MABC have a clear direction, and convergence or divergence of the objective function in the search space can be determined quickly. This method is similar to the roulette way of nectar source for the onlooker bees, and so sensitivity and pheromone option instead of roulette way can be applied.
3.4. Updating Step Length with Exponential Function
According to solution search equation (1) of ABC, the new candidate solution is generated through moving the old one towards another selected random solution of the population. This method can increase the probability of locating global optimal solution, but it cannot guarantee obtaining a better solution. Hence, the convergence speed is slow. If each new candidate solution is better than the old one, then convergence speed will become faster. However, if only the better solution is considered, then the algorithm could get trapped into local optimum [32]. In ABC, the updating step length denoted by (i.e., in (1)) plays a very important role and controls the ability to search (local or global) and the convergence rate. On the other hand, if the value of is too large, then the algorithm is easy to lose the global optimal solution. And, on the contrary, the algorithm is easy to show the phenomenon of stagnation and fall into local optimum [33]. Therefore, it is very necessary to design a new suitable step length associated with the performance of the algorithm.
The following exponential function is used in this paper to update step length:and the improved population moving step length isfor and , where is the same as in (1). Thus, the location of the updated solution isand the location of the random generated solutions iswhere and are the same as in .
Now one can know that with the increase of evolution algebra, the updating step will be increased and become larger and larger. In the early stage of the evolution, the improved evolution algebra can be conducted a comprehensive local search, and make the algorithm better position to the target area. In the middle and later periods of the evolution, the improved evolution algebra will be carried out global search and can speed up the convergence rate so as to reach to optimal solution.
3.5. Flowchart of MABC
To clearly know MABC, the detailed flowchart of MABC is shown in Algorithm 3, where the max iteration as the termination criteria is adopted the same as in Algorithm 1, and CS denotes the number of colony size, which is equal to the sum of the numbers of employed bees and onlooker bees.

4. Simulation and Analysis
In this section, the performance and accuracy of MABC, ABC, NABC, and DEABC will be analyzed and compared through the figures, the total run time, mean value, and standard deviation of six benchmark functions from the new test function versions CEC13, respectively.
4.1. Convergence Performance Analysis
In order to verify the validity of MABC, six benchmark functions listed in Table 3 (see [11, 27]) are used in the experiments. According to their characteristics, these six functions are divided into two groups. The Sphere, SumSquares, and Schwefel 2.22 are unimodal functions, and Ackley, Griewank, and Rastrigin are multimodal functions. For each benchmark function, two dimension sizes, 20 and 40, of solution space are tested. The colony size is 80 and the max iteration is 1500; limit is 100. Each of the experiments is repeated 10 times.

In order to further compare the performance of four proposed algorithms, convergence performance of the six functions for ABC, DEABC, NABC, and MABC is plotted to dimension sizes 20 and 40, respectively. See Figures 1–6.
From the convergence performances of SumSquares, Griewank, and Rastrigin functions, one can see that when , ABC converges with the slowest speed and locates a local optimal solution. DEABC has the fastest speed at the beginning and goes into the global optimal solution. NABC also goes into the global optimal solution, but the speed is slower than those of DEABC and MABC. Further, MABC has the fastest speed at the end and is the first one to attain the true global number. When , ABC and DEABC fall into the global solution. NABC has the fastest speed at the beginning and attains the global optimal solution at the end, but the true global number of NABC is smaller than that of MABC. MABC goes into the global optimal solution at the end, and the speed is faster than that of NABC.
The convergence performance of Sphere function exhibits that various behavior for both dimension sizes. When , ABC converges with the slowest speed, and NABC and DEABC have fastest speed in the first place. Later on, MABC is the first to arrive to the global optimal solution, the second one is DEABC, and the third one is NABC. And ABC locates a local optimal solution. When , ABC also goes into the local optimal. NABC, DEABC, and MABC have the similar speed originally; hereafter MABC has quicker speed than NABC and DEABC. Finally, these three algorithms come to the global optimal solution, but the best one is MABC, the second one is DEABC, and the last one is NABC.
The convergence performances of Schwefel 2.22 and Ackley functions demonstrate that, for both dimension sizes of 20 and 40, ABC, NABC, and DEABC all fall into the global optimal solution, but DEABC converges with the fastest speed originally. Later on, MABC has the fastest speed and locates the global optimal solution at the end.
From Figures 1–6, it can be comprehensively deduced that ABC converges with the slowest speed and tends to a local optimal solution. DEABC and NABC have faster speed than ABC but sometimes enter into local optimal solution. MABC has fastest speed at the end and achieves the really global optimal solution.
4.2. Comparison of Mean Value and Standard Deviation
The results of comparing the mean value and standard deviation are shown in Table 4, where the first column shows the benchmark functions, the “” column represents the dimension of function, the “Mean” column contains the average best values of benchmark functions, and the “SD” column shows the standard deviation of the best value.

For six benchmark functions, one can know that when , the mean values of ABC, NABC, DEABC, and MABC are better than those of them when . For SumSquares, Griewank, and Rastrigin functions, it is observed that when , the mean value of NABC has better mean than that of DEABC. On the other hand, in terms of standard deviation, ABC and NABC are less than DEABC for Sphere and Ackley functions. In addition, MABC has better average best values than DEABC, and DEABC has better average best values than ABC and NABC. The least standard deviation of MABC to six test functions with the dimensions and is , and the standard deviations of MABC are always less than DEABC; meanwhile the standard deviation of DEABC is less than ABC and NABC.
Hence, by the results of six test functions with the dimensions and , MABC have better results than ABC, NABC, and DEABC. For all test functions, the mean and standard deviation of MABC results are several orders of magnitude better than ABC, NABC, and DEABC, particularly in case of SumSquares, Schwefel 2.22, and Ackley functions. Therefore, MABC has more accurate solution and better stability than ABC, NABC, and DEABC.
4.3. Comparison of Total Run Time
This subsection will show effectively that the searching speed of MABC is fastest through the phenomenon for total run time of solutions (see Figure 7). To all the functions, when or 40, NABC costs the more time, ABC and DEABC follow closely, MABC spends time the least, and the time of search solution of MABC is independent of dimension sizes. And the Sphere function costs least computational time, while the Rastrigin function takes the maximum one, so it shows that the finding solution time is related to the function types. Further, the time of search solution of MABC is less than that of any one of ABC, NABC, and DEABC.
Moreover, based on the six benchmark functions with the dimension or , Table 5 displays all of the reduction rates, and the reduction rate of total iteration time is at least 73.52% and up to 85.88% for MABC to DEABC, MABC to NABC, and MABC to ABC.

5. Concluding Remarks
In this paper, to improve classical artificial bee colony algorithms, a new modified artificial bee colony (MABC) algorithm was developed, which is poor at exploitation and has some defects such as slow search speed, poor population diversity, the stagnation in the working process, and being trapped into the local optimal solution.
In order to make the initial population diversity distribute evenly and to take full use of the search space information, the initial population was constructed by using the oppositionbased learning theory. Next, a new type method of grouping was applied to reduce the difference between each group and replaced a greedy method (i.e., the roulette wheel way) by free search algorithmsensitivity with pheromone. Then, a suitable exponential function updating length step was designed for replacing the original random step and carrying out faster and more stable searching of the global optimal solution.
Finally, from the new test function versions CEC13, six benchmark functions with the dimensions and were chosen and applied in the experiments for analyzing and comparing the iteration speed and accuracy of MABC to differential evolution artificial bee colony algorithm (DEABC), novel artificial bee colony algorithm (NABC), and basic artificial bee colony algorithm (ABC). With respect to the six benchmark functions with the dimensions and , respectively, one can see from the experimental results in Tables 4 and 5 that the smallest standard deviation of the new modified algorithm is , and the reduction rate of total iteration time for the new modified algorithm relative to the other three proposed artificial bee colony algorithms is at least 73.52% and up to 85.88%. Moreover, the experimental results presented in Figures 1–7 show that MABC has faster and more stable searching and can increase poor population diversity and avoid falling into the local optimal solutions.
MABC is superior to ABC, NABC, and DEABC, and the feasibility and effectiveness of MABC are very easy to be implemented. However, some other stateoftheart algorithms may be selected as comparable algorithms, and more complex test functions and more diverse set of benchmark functions may be implemented to validate the benefits of MABC. Thus, the following works as open questions will be worth further studying:(i)Some applications of MABC in the areas of pattern classification, fuzzy control, nonlinear system modeling, parameter tuning of proportionalintegralderivative controller, and so on.(ii)Selecting more stateoftheart algorithms as comparable algorithms.(iii)Simulating more diverse set of benchmark functions for corresponding swarm intelligence algorithms.(iv)Implementing more complex functions for validating the benefits of MABC and other corresponding improving swarm intelligence algorithms.
Conflict of Interests
The authors declare that they have no competing interests regarding the publication of this paper.
Acknowledgments
This work was partially supported by Sichuan Province Cultivation Fund Project of Academic and Technical Leaders and the Open Research Fund of Key Laboratory of Higher Education of Sichuan Province for Enterprise Informationalization and Internet of Things (2013WZJ01) and the Scientific Research Project of Sichuan University of Science & Engineering (2015RC07) and cofinanced by National Natural Science Foundation of China (61573010 and 11501391).
References
 W. L. Xiang, S. F. Ma, and M. Q. An, “Habcde: a hybrid evolutionary algorithm based on artificial bee colony algorithm and differential evolution,” Applied Mathematics and Computation, vol. 238, pp. 370–386, 2014. View at: Publisher Site  Google Scholar  MathSciNet
 H.C. Tsai, “Integrating the artificial bee colony and bees algorithm to face constrained optimization problems,” Information Sciences, vol. 258, pp. 80–93, 2014. View at: Publisher Site  Google Scholar  MathSciNet
 M. F. Tasgetiren, Q.K. Pan, P. N. Suganthan, and A. Oner, “A discrete artificial bee colony algorithm for the noidle permutation flowshop scheduling problem with the total tardiness criterion,” Applied Mathematical Modelling, vol. 37, no. 1011, pp. 6758–6779, 2013. View at: Publisher Site  Google Scholar  MathSciNet
 K. S. Tang, K. F. Man, S. Kwong, and Q. He, “Genetic algorithms and their applications,” IEEE Signal Processing Magazine, vol. 13, no. 6, pp. 22–37, 1996. View at: Publisher Site  Google Scholar
 J. Kennedy, “Particle swarm optimization,” in Encyclopedia of Machine Learning, C. Sammut and G. I. Webb, Eds., pp. 760–766, Springer US, 2010. View at: Google Scholar
 M. Dorigo and L. M. Gambardella, “Ant colony system: a cooperative learning approach to the traveling salesman problem,” IEEE Transactions on Evolutionary Computation, vol. 1, no. 1, pp. 53–66, 1997. View at: Publisher Site  Google Scholar
 D. Simon, “Biogeographybased optimization,” IEEE Transactions on Evolutionary Computation, vol. 12, no. 6, pp. 702–713, 2008. View at: Publisher Site  Google Scholar
 S. Zhang and S. Liu, “A novel artificial bee colony algorithm for function optimization,” Mathematical Problems in Engineering, vol. 2015, Article ID 129271, 10 pages, 2015. View at: Publisher Site  Google Scholar
 D. Karaboga, “An idea based on honey bee swarm for numerical optimization,” Technical Report tr06, Computer Engineering Department, Engineering Faculty, Erciyes University, Kayseri, Turkey, 2005. View at: Google Scholar
 M. ElAbd, “Performance assessment of foraging algorithms vs. evolutionary algorithms,” Information Sciences, vol. 182, no. 1, pp. 243–263, 2012. View at: Publisher Site  Google Scholar  MathSciNet
 T. Y. Lim, M. A. AlBetar, and A. T. Khader, “Adaptive pair bonds in genetic algorithm: an application to realparameter optimization,” Applied Mathematics and Computation, vol. 252, pp. 503–519, 2015. View at: Publisher Site  Google Scholar
 Y. Xiang, Y. Peng, Y. Zhong, Z. Chen, X. Lu, and X. Zhong, “A particle swarm inspired multielitist artificial bee colony algorithm for realparameter optimization,” Computational Optimization and Applications, vol. 57, no. 2, pp. 493–516, 2014. View at: Publisher Site  Google Scholar
 D. Karaboga, B. Gorkemli, C. Ozturk, and N. Karaboga, “A comprehensive survey: artificial bee colony (ABC) algorithm and applications,” Artificial Intelligence Review, vol. 42, no. 1, pp. 21–57, 2014. View at: Publisher Site  Google Scholar
 J. C. Bansal, H. Sharma, and S. S. Jadon, “Artificial bee colony algorithm: a survey,” International Journal of Advanced Intelligence Paradigms, vol. 5, no. 12, pp. 123–159, 2013. View at: Publisher Site  Google Scholar
 A. L. Bolaji, A. T. Khader, M. A. AlBetar, and M. A. Awadallah, “Artificial bee colony algorithm, its variants and applications: a survey,” Journal of Theoretical and Applied Information Technology, vol. 47, no. 2, pp. 434–459, 2013. View at: Google Scholar
 H. Habbi, Y. Boudouaoui, D. Karaboga, and C. Ozturk, “Selfgenerated fuzzy systems design using artificial bee colony optimization,” Information Sciences, vol. 295, pp. 145–159, 2015. View at: Publisher Site  Google Scholar  MathSciNet
 L. N. Vitorino, S. F. Ribeiro, and C. J. A. BastosFilho, “A mechanism based on artificial bee colony to generate diversity in particle swarm optimization,” Neurocomputing, vol. 148, pp. 39–45, 2015. View at: Publisher Site  Google Scholar
 M. S. Kiran and O. Findik, “A directed artificial bee colony algorithm,” Applied Soft Computing Journal, vol. 26, pp. 454–462, 2015. View at: Publisher Site  Google Scholar
 D. Karaboga and B. Gorkemli, “A quick artificial bee colony (qABC) algorithm and its performance on optimization problems,” Applied Soft Computing Journal, vol. 23, pp. 227–238, 2014. View at: Publisher Site  Google Scholar
 N. Imanian, M. E. Shiri, and P. Moradi, “Velocity based artificial bee colony algorithm for high dimensional continuous optimization problems,” Engineering Applications of Artificial Intelligence, vol. 36, pp. 148–163, 2014. View at: Publisher Site  Google Scholar
 H. Wang, Z. Wu, S. Rahnamayan, H. Sun, Y. Liu, and J.s. Pan, “Multistrategy ensemble artificial bee colony algorithm,” Information Sciences, vol. 279, pp. 587–603, 2014. View at: Publisher Site  Google Scholar  MathSciNet
 W.F. Gao, S.Y. Liu, and L.L. Huang, “Enhancing artificial bee colony algorithm using more informationbased search equations,” Information Sciences, vol. 270, pp. 112–133, 2014. View at: Publisher Site  Google Scholar  MathSciNet
 S. Das, S. Biswas, and S. Kundu, “Synergizing fitness learning with proximitybased food source selection in artificial bee colony algorithm for numerical optimization,” Applied Soft Computing, vol. 13, no. 12, pp. 4676–4694, 2013. View at: Publisher Site  Google Scholar
 M. X. Zang, X. Ma, and Y. M. Duan, “Improved artificial bee colony algorithm,” Journal of Xidian University, vol. 42, no. 2, pp. 65–70, 2015. View at: Google Scholar
 C. Zhang, J. Zheng, and Y. Zhou, “Two modified artificial bee colony algorithms inspired by grenade explosion method,” Neurocomputing, vol. 151, no. 3, pp. 1198–1207, 2015. View at: Publisher Site  Google Scholar
 M. Mernik, S.H. Liu, D. Karaboga, and M. Črepinšek, “On clarifying misconceptions when comparing variants of the artificial bee colony algorithm by offering a new implementation,” Information Sciences, vol. 291, pp. 115–127, 2015. View at: Publisher Site  Google Scholar  MathSciNet
 M. S. Kiran, H. Hakli, M. Gunduz, and H. Uguz, “Artificial bee colony algorithm with variable search strategy for continuous optimization,” Information Sciences, vol. 300, pp. 140–157, 2015. View at: Publisher Site  Google Scholar  MathSciNet
 T. G. Chen and R. B. Xiao, “Modeling design iteration in product design and development and its solution by a novel artificial bee colony algorithm,” Computational Intelligence and Neuroscience, vol. 2014, Article ID 240828, 13 pages, 2014. View at: Publisher Site  Google Scholar
 B. Li, “Research on WNN modeling for gold price forecasting based on improved artificial bee colony algorithm,” Computational Intelligence and Neuroscience, vol. 2014, Article ID 270658, 10 pages, 2014. View at: Publisher Site  Google Scholar
 S. Wei and N. Z. Lu, “An improvement for artificial bee algorithm,” Modern Computer, vol. 6, no. 17, pp. 25–29, 2014. View at: Google Scholar
 L. Li, F. M. Yao, L. J. Tan et al., “A novel DEABCbased hybrid algorithm for global optimization,” in Proceedings of the 7th International Conference on Intelligent Computing (ICIC '11), Zhengzhou, China, August 2011, D. S. Huang, Y. Gan, P. Premaratne, and K. Han, Eds., Lecture Notes in Computer Science, pp. 558–565, Springer, Berlin, Germany, 2012. View at: Publisher Site  Google Scholar
 S.M. Chen, A. Sarosh, and Y.F. Dong, “Simulated annealing based artificial bee colony algorithm for global numerical optimization,” Applied Mathematics and Computation, vol. 219, no. 8, pp. 3575–3589, 2012. View at: Publisher Site  Google Scholar  MathSciNet
 S. C. Satapathy and A. Naik, “Modified teachinglearningbased optimization algorithm for global numerical optimization—a comparative study,” Swarm and Evolutionary Computation, vol. 16, pp. 28–37, 2014. View at: Publisher Site  Google Scholar
 Q. Xu, L. Wang, N. Wang, X. Hei, and L. Zhao, “A review of oppositionbased learning from 2005 to 2012,” Engineering Applications of Artificial Intelligence, vol. 29, pp. 1–12, 2014. View at: Publisher Site  Google Scholar
 K. Penev, “Free search—comparative analysis 100,” International Journal of Metaheuristics, vol. 3, no. 2, pp. 118–132, 2014. View at: Publisher Site  Google Scholar
Copyright
Copyright © 2016 Wei Mao et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.