Research Article  Open Access
OppositionBased Animal Migration Optimization
Abstract
AMO is a simple and efficient optimization algorithm which is inspired by animal migration behavior. However, as most optimization algorithms, it suffers from premature convergence and often falls into local optima. This paper presents an oppositionbased AMO algorithm. It employs oppositionbased learning for population initialization and evolution to enlarge the search space, accelerate convergence rate, and improve search ability. A set of wellknown benchmark functions is employed for experimental verification, and the results show clearly that oppositionbased learning can improve the performance of AMO.
1. Introduction
Many real world problems can be summarized as optimization problems. Optimization problems play an important role in both industrial application and scientific research. In the past decades, different optimization algorithms have been proposed. Among them, genetic algorithm may be the first and popular algorithm inspired by natural genetic variation and natural selection [1, 2]. Inspired by the social behavior of bird flocking or fish school, particle swarm algorithm was developed by Kennedy and Eberhart in 1995 [3, 4]. Artificial bee colony algorithm was proposed by Karaboga and Basturk in 2005, which simulates the foraging behavior of bee swarm [5, 6]. Ant colony optimization (ACO) which simulates the action of ants was first introduced by Dorigo [7, 8]. Animal Migration Optimization (AMO) as a new optimization algorithm inspired by animal migration behavior was first proposed by Li et al. in [9]. AMO simulates the widespread migration phenomenon in the animal kingdom, through the change of position, replacement of individual, and finding the optimal solution gradually. AMO has obtained good experimental results on many optimization problems.
Optimization algorithms often begin from an initial set of variables which are generated randomly and through iteration to obtain the global optimal solutions or the maximum of the objective function. It is known that the performance of these algorithms is highly related to diversity of particles; it may easily fall into local optima and has slow convergence rate and poor accuracy in the later stage of evolution. In recent years, many efforts have been done to improve the performance of different algorithms.
The concept of oppositionbased learning (OBL) was first introduced by Tizhoosh [10]. The main idea is to consider current candidate solution and its opposite candidate in order to enlarge the search scope, and it also uses elite selection mechanism to speed up the convergence speed and find the optimal solution. It has been proved in [11], an opposite learning mechanism has more chance to be closer to the global optimum solution than a random candidate solution. The oppositionbased learning idea has been successfully applied on GA [11] that PSO [12–16], DE [17, 18], artificial neural networks (ANN) [19, 20], and ant colony optimization (ACO) [21]; experimental results show that oppositionbased learning can improve the search capabilities of the algorithm to some extent.
This paper presents an algorithm to improve the performance of AMO. In the absence of priori information about the population, we employ oppositionbased learning during population initialization and population evolution. Through the introduction of oppositionbased learning mechanism, it can transform solutions from current search space to a new search space to enlarge the search space. By means of selecting the better solution between current solution and the opposite solution, it will improve search ability and accelerate convergence rate, and it has more chance to find the global optima.
The rest of this paper is organized as follows. Section 2 briefly introduces the animal migration optimization algorithm. Section 3 gives a simple description of oppositionbased learning. Section 4 explains an implementation of the proposed algorithm, oppositionbased AMO algorithm. Section 5 presents a comparative study among AMO, OPAMO, and other optimization algorithms on 23 benchmark problems. Finally, the work is concluded in Section 6.
2. Animal Migration Optimization
AMO is a new heuristic optimization algorithm inspired by the behavior of animal migration which is a ubiquitous phenomenon that can be found in all major animal groups, such as birds, mammals, fish, reptiles, amphibians, insects, and crustaceans [9].
In this algorithm, there are mainly two processes. In the first process the algorithm simulates how the groups of animals move from current position to a new position. During this process, each individual should obey three main rules: move in the same direction as its neighbors; remain close to its neighbors; avoid collisions with its neighbors. We select one neighbor randomly and update the position of the individual according to this neighbor, as can be seen in the following formula: where is the current position of the neighborhood, is produced using a random number generator controlled by a Gaussian distribution, is the current position of th individual, and is the new position of th individual.
In the following process, the algorithm simulates how new animals are introduced to the group during the migration. During the population updating process, some animals will leave the group and some new animals will join the new population. We assume that the number of available animals is fixed. The animals will be replaced by some new individuals with a probability Pa. The probability is used according to the quality of the fitness. For the best fitness, the probability Pa is 1/NP. For the worst fitness, the probability is 1. This process can be shown in Algorithm 1, where are randomly chosen integers and. After producing a new solution, it will be evaluated and compared with . If the objective fitness of is smaller than the fitness of , is accepted as a new basic solution; otherwise, would be obtained.

To verify the performance of AMO, 23 benchmark functions chosen from the literature are employed. The results show that the proposed algorithm clearly outperforms some evolution algorithms from the literature.
3. OppositionBased Learning
In the evolutionary algorithm, algorithm often starts from random initial population until a satisfactory solution is found. If no population prior information is known, the speed of evolution is relation to the distance between the initial particles and the best particle. If we select some initial particles close to the best individual, we can accelerate the convergence of the algorithm to some extent.
The opposition learning algorithm is a new type of reinforcement learning algorithm, and it has been proven to be an effective concept to enhance various optimization approaches [22]. This algorithm utilizes opposition learning mechanism to generate opposite population and employs elite selection to choose the individual closer to the best individual as the member of initial population, thus facilitating the overall evolutionary convergence speed.
For ease of description, we give the definition of the opposition point first.
Definition 1. Let be a real number; the opposite point is defined as
Similarly, the definition can be extended to high dimensional space.
Definition 2. Let as a point in dimension space, where, for all ; the point is defined as the opposition value of , in which
Through the two previous definitions, we can conclude that the opposite learning algorithm is defined as follows.
Definition 3. is a candidate solution in dimension space; assume is the fitness function which is used to evaluate the fitness of the candidate. According to the definition of the opposite point, is the opposite of . If is better than , then choose as candidate instead of ; otherwise, is obtained, through evaluating the fitness of and the opposite to get the better individual.
4. Enhanced AMO Using OppositionBased Learning
4.1. Basic Concept
According to the probability principle, each randomly generated candidate solution has fifty percent probability chance away from the optimal solution compared to its opposition solution; AMO algorithm usually generates candidate solutions randomly in the search space as initial population because of the lack of prior information. During the optimization process, the calculation time of finding optimal solution will be changed according to the distance between the candidate solution and the optimal solution. By applying oppositionbased learning, we not only evaluate the current candidate but also calculate its opposition candidate ; this will provide more chance of finding the solution closer to the global optimal value.
Let be a solution in current search space, ; the new solution in the opposite space is
From the above definition, we can infer. Obviously, through the opposite transformation, the center of search space is changed from to .
Let , where is a real number in , and it is often generated randomly; this denotes the probability of reverse learning.
4.2. OppositionBased AMO Algorithm
In this paper, we introduce a novel algorithm which combines opposition learning algorithm with AMO algorithm. Through the introduction of opposition learning mechanism, we consider the candidate solution in both current search space and its opposite space simultaneously. By selecting the better individuals in the two search spaces, it can provide more chance to find the global optimal solution and largely speed up the convergence.
During the process of population initialization and evolution, through the introduction of opposition learning mechanism, more candidate solutions will be considered, and we choose the most likely candidate solutions for evolution.
We described the oppositionbased population initial process in detail.(1)Initialize the population in the search space randomly.(2)Calculate opposition population according to initial population; each dimension is calculated as follows: where and denote the th variable of the th vector of the population and opposite population, respectively, and is a real number in which denotes the probability of reverse learning.(3)Choose fittest individual as initial population from the union of random population and opposition population according to the value of fitness.
During the evolution process, we still adopt opposition learning method to increase the opportunity of finding the optimal solution. When a new individual is generated or joined, its opposition value is considered; if the fitness of opposite solution is better than the new individual, the opposite solution is adopted; otherwise, the new individual is obtained.
However, opposition learning method could not be suitable for all kinds of optimization problems. For instance, when calculating the opposition candidate, the solution may jump away from the solution space. If this happens, the solution will be invalid; to avoid this case, the transformed candidate is assigned to a random value as follows: where rand is a random number between and .
The opposition learning based AMO algorithm is described in Algorithm 2.

5. Experimental Results
To evaluate the performance of our algorithm, we applied it to 23 standards benchmark functions as shown in Table 1. These functions have been widely used in the literature.

The maximum numbers of generations are 1500 for , , , , and , 2000 for and , 3000 for , , and , and 5000 for , and, 400 for , 100 for , , , , , , and , 30 for , and 200 for . The population size is 50 because this algorithm has two phases. The results of the algorithm on the 23 test problems are presented in Tables 2, 3, and 4.



As seen from Table 2 to Table 4, oppositionbased AMO algorithm significantly improves the results on functions , , , , , and and achieves the same performance with AMO on most other functions, and it is better than PSO, DE, ABC, and FA. This denotes that the algorithm can accelerate convergence rate and find better solutions for some optimization problems, and it did not lead to the premature of algorithm.
Figure 1 shows the evolutionary process of test function by AMO and opposition learning based AMO. The horizontal ordinate denotes the evolution iterations, and the vertical ordinate is the optimal value of objective function. We can see from the graph that the convergence rate of oppositionbased AMO is obviously accelerated and the accuracy of optimization is also enhanced.
6. Conclusions
A novel oppositionbased AMO algorithm is proposed in this paper. This approach can provide more chance to find better solutions by transforming candidate solutions from current search space to a new search space. Experimental results show that, compared with previous AMO, the proposed algorithm is efficient in most of the test functions. However, we can also see from the experimental results that this algorithm is not suitable for all kinds of problems; for some optimization problems, algorithm has no significant improvement. How to improve the algorithm to adapt to a more optimization problem is worth of further study.
Conflict of Interests
The authors declare that they do not have any commercial or associative interest that represents a conflict of interests in connection with the work submitted.
References
 M. Melanie, An Introduction to Genetic Algorithms, MIT Press, Cambridge, Mass, USA, 1999.
 S. N. Sivanandam and S. N. Deepa, Introduction to Genetic Algorithms, Springer, Berlin, 2008. View at: MathSciNet
 J. Kennedy and R. Eberhart, “Particle swarm optimization,” in Proceedings of the IEEE International Conference on Neural Networks, vol. 14, pp. 1942–1948, December 1995. View at: Google Scholar
 A. P. Engelbrecht, Fundamentals of Computational Swarm Intelligence, John Wiley & Sons, Hoboken, NJ, USA, 2005. View at: Zentralblatt MATH
 D. Karaboga and B. Basturk, “A powerful and efficient algorithm for numerical function optimization: artificial bee colony (ABC) algorithm,” Journal of Global Optimization, vol. 39, no. 3, pp. 459–471, 2007. View at: Publisher Site  Google Scholar  Zentralblatt MATH  MathSciNet
 D. Karaboga and B. Basturk, “On the performance of artificial bee colony (ABC) algorithm,” Applied Soft Computing Journal, vol. 8, no. 1, pp. 687–697, 2008. View at: Publisher Site  Google Scholar
 M. Dorigo, Optimization, learning and natural algorithms [Ph.D. dissertation], Politecnico di Milano, Milano, Italy, 1992.
 M.H. Lin, J.F. Tsai, and L.Y. Lee, “Ant colony optimization for social utility maximization in a multiuser communication system,” Mathematical Problems in Engineering, vol. 2013, Article ID 798631, 8 pages, 2013. View at: Publisher Site  Google Scholar
 X. Li, J. Zhang, and M. Yin, “Animal migration optimization: an optimization algorithm inspired by animal migration behavior,” Neural Computing and Applications, 2013. View at: Publisher Site  Google Scholar
 H. R. Tizhoosh, “Oppositionbased learning: a new scheme for machine intelligence,” in International Conference on Computational Intelligence for Modelling, Control and Automation (CIMCA '05), pp. 695–701, November 2005. View at: Google Scholar
 S. Rahnamayan, H. R. Tizhoosh, and M. M. A. Salama, “Opposition versus randomness in soft computing techniques,” Applied Soft Computing Journal, vol. 8, no. 2, pp. 906–918, 2008. View at: Publisher Site  Google Scholar
 Z. Lin and L. Wang, “A new oppositionbased compact genetic algorithm with fluctuation,” Journal of Computational Information Systems, vol. 6, no. 3, pp. 897–904, 2010. View at: Google Scholar
 H. Lin and H. Xingshi, “A novel oppositionbased particle swarm optimization for noisy problems,” in Proceedings of the 3rd International Conference on Natural Computation (ICNC '07), pp. 624–629, Haikou, China, August 2007. View at: Publisher Site  Google Scholar
 H. Wang, H. Li, Y. Liu, C. Li, and S. Zeng, “Oppositionbased particle swarm algorithm with Cauchy mutation,” in Proceedings of the 2007 IEEE Congress on Evolutionary Computation (CEC '07), pp. 4750–4756, Singapore, September 2007. View at: Publisher Site  Google Scholar
 H. Wang, Z. Wu, S. Rahnamayan, Y. Liu, and M. Ventresca, “Enhancing particle swarm optimization using generalized oppositionbased learning,” Information Sciences, vol. 181, no. 20, pp. 4699–4714, 2011. View at: Publisher Site  Google Scholar  MathSciNet
 H. Wang, Z. Wu, S. Rahnamayan, and J. Wang, “Diversity analysis of oppositionbased differential evolutionan experimental study,” in Proceedings of the International Symposium on Intelligence Computation and Applications, pp. 95–102, 2010. View at: Google Scholar
 S. Rahnamayan, H. R. Tizhoosh, and M. M. A. Salama, “Oppositionbased differential evolution algorithms,” in Proceedings of the IEEE Congress on Evolutionary Computation (CEC '06), pp. 2010–2017, Vancouver, Canada, July 2006. View at: Google Scholar
 R. S. Rahnamayan, H. R. Tizhoosh, and M. M. A. Salama, “Oppositionbased differential evolution,” IEEE Transactions on Evolutionary Computation, vol. 12, no. 1, pp. 64–79, 2008. View at: Publisher Site  Google Scholar
 M. Ventresca and H. R. Tizhoosh, “Improving the convergence of backpropagation by opposite transfer functions,” in Proceedings of the International Joint Conference on Neural Networks (IJCNN '06), pp. 4777–4784, Vancouver, Canada, July 2006. View at: Google Scholar
 M. Ventresca and H. R. Tizhoosh, “Opposite transfer functions and backpropagation through time,” in Proceedings of the IEEE Symposium on Foundations of Computational Intelligence (FOCI '07), pp. 570–577, Honolulu, Hawaii, USA, April 2007. View at: Publisher Site  Google Scholar
 A. R. Malisia and H. R. Tizhoosh, “Applying oppositionbased ideas to the Ant Colony System,” in Proceedings of the IEEE Swarm Intelligence Symposium (SIS '07), pp. 182–189, Honolulu, Hawaii, USA, April 2007. View at: Publisher Site  Google Scholar
 W.f. Gao, S.y. Liu, and L.l. Huang, “Particle swarm optimization with chaotic oppositionbased population initialization and stochastic search technique,” Communications in Nonlinear Science and Numerical Simulation, vol. 17, no. 11, pp. 4316–4327, 2012. View at: Publisher Site  Google Scholar  Zentralblatt MATH  MathSciNet
Copyright
Copyright © 2013 Yi Cao et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.