Mathematical Problems in Engineering

Volume 2012 (2012), Article ID 761708, 12 pages

http://dx.doi.org/10.1155/2012/761708

## Opposition-Based Barebones Particle Swarm for Constrained Nonlinear Optimization Problems

School of Information Engineering, Nanchang Institute of Technology, Nanchang 330099, China

Received 29 December 2011; Revised 9 April 2012; Accepted 10 May 2012

Academic Editor: Jianming Shi

Copyright © 2012 Hui Wang. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

#### Abstract

This paper presents a modified barebones particle swarm optimization (OBPSO) to solve constrained nonlinear optimization problems. The proposed approach OBPSO combines barebones particle swarm optimization (BPSO) and opposition-based learning (OBL) to improve the quality of solutions. A novel boundary search strategy is used to approach the boundary between the feasible and infeasible search region. Moreover, an adaptive penalty method is employed to handle constraints. To verify the performance of OBPSO, a set of well-known constrained benchmark functions is used in the experiments. Simulation results show that our approach achieves a promising performance.

#### 1. Introduction

Many engineering problems can be converted to constrained optimization problems. The aim of constrained optimization is to find a feasible solution with minimized cost (In this paper, we only consider minimization problems). A general constrained minimization optimization problem can be defined as follows. subject to where is inequality constraint, is the equality constraint, is the number of constraints, is the number of inequality constraints, and is the number of equality constraints.

Particle swarm optimization (PSO) is a population-based stochastic search algorithm developed by Kennedy and Eberhart [1]. Although PSO shares many similarities with evolutionary algorithms (EAs), the standard PSO does not use evolution operators such as crossover and mutation. For PSO's simple concept, easy implementation, and effectiveness, it has been widely applied to many optimization areas.

Although PSO has shown a good performance over many optimization problems, it does not work well when solving complex problems. Especially for constrained optimization problems, the standard PSO could hardly search promising solutions. The possible reason is that constrained optimization problems are usually multimodal and having some constraints. PSO could easily fall into local minima and hardly search feasible solutions. To enhance the performance of PSO on constrained optimization problems, many improved PSO variants have been proposed in the past several years.

Meng et al. [2] used a quantum-inspired PSO (QPSO) to solve constrained economic load dispatch. The QPSO shows stronger search ability and quicker convergence speed, not only because of the introduction of quantum computing theory, but also due to two-novel strategies: self-adaptive probability selection and chaotic sequences mutation. Coelho [3] presented a novel quantum-behaved PSO (QPSO) to solve constrained engineering design problems. The proposed approach embedded a Gaussian mutation into QPSO to prevent premature convergence to local minima. Sun et al. [4] proposed an improved vector PSO (IVPSO) based on multidimensional search, in which a simple constraint-preserving method is used to handle constraints. Liu et al. [5] used a hybrid PSO called PSO-DE to solve constrained numerical and engineering optimization problems. The PSO-DE integrates PSO with differential evolution (DE) to obtain a good performance. Venter and Haftka [6] proposed a new method to solve constrained optimization problems. The constrained, single objective optimization problem is converted into an unconstrained, biobjective optimization problem that is solved using a multiobjective PSO algorithm. Lu and Chen [7] presented an enhanced PSO by employing a dynamic inertia weight to avoid premature convergence. The inertia weight of every individual is dynamically controlled by the Euclidean distance between individual and the global best individual. Daneshyari and Yen [8] proposed a cultural-based constrained PSO to incorporate the information of the objective function and constraint violation. The archived information facilitates communication among swarms in the population space and assists in selecting the leading particles in three different levels: personal, swarm, and global levels.

There have been many modifications to the original PSO algorithm to improve the efficiency and robustness of the search. Although these modified PSO variants have shown good search abilities, their performance greatly depends on the control parameters in the velocity updating model, such as inertia weight () and acceleration coefficients ( and ). Recently, a parameter-free PSO, known barebones PSO (BPSO) [9], used Gaussian normal distribution to update the particles in the population. It does not involve inertia weight, acceleration coefficients, and velocity. Its performance has been found to be competitive, and a number of BPSO algorithms have been proposed in the past several years. Omran et al. [10] incorporate the idea of BPSO into DE. Krohling and Mendel [11] employed a jump strategy in BPSO to avoid premature convergence. Motivated by the idea of BPSO, this paper presents an improved BPSO, namely OBPSO, to solve constrained nonlinear optimization problems. In OBPSO, opposition-based learning (OBL) concept [12] is used for population initialization and generation jumping. To verify the performance of OBPSO, a set of well-known constrained benchmark problems are used in the experiments. Results obtained by the proposed OBPSO are compared with those in the literature and discussed.

The rest of the paper is organized as follows. In Section 2, the standard PSO and barebones PSO are briefly introduced. Section 3 describes our proposed approach. Section 4 presents experimental simulations, results, and discussions. Finally, the work is concluded in Section 5.

#### 2. Belief Descriptions of PSO and Barebones PSO

In traditional PSO, a member in the swarm, called a particle, represents a potential solution in the -dimensional search space. Each particle has two vectors: velocity and position. It is attracted by its previous best particle () and the global best particle () during the evolutionary process. The velocity and position of the th dimension of the th particle are updated according to (2.1) [13]: where is the particle's index, is the population size, is the position of the th particle; represents the velocity of the th particle; the is the best previous position yielding the best fitness value for the th particle; is the global best particle found by all particles so far. The parameter , called inertia factor, which is used to balance the global and local search abilities of particles [13], rand and rand are two random numbers generated independently within the range of , and are two learning factors which control the influence of the social and cognitive components, and indicates the iteration number.

Recently, Kennedy [9] developed the barebones PSO (BPSO). This new version of PSO eliminates the velocity term, and the position is updated as follows. where is the position of the th particle in the population, and represents a Gaussian distribution with mean and standard deviation .

Note that the particle positions are sampled by the above Gaussian distribution. The BPSO facilitates initial exploration, due to large deviation (initially, will be far from the ). As the number of generation increases, the deviation approaches to zero, by focussing on exploitation of the and [14].

#### 3. Opposition-Based Barebones PSO (OBPSO)

##### 3.1. Opposition-Based Learning

Opposition-based learning (OBL) developed by Tizhoosh [12] is a new concept in computational intelligence. It has been proven to be an effective concept to enhance various optimization approaches [15–17]. When evaluating a solution to a given problem, simultaneously computing its opposite solution will provide another chance for finding a candidate solution which is closer to the global optimum.

Let be a candidate solution in a -dimensional space, where and , . The opposite solution is defined by [15]

By staying within variables’ interval static boundaries, we would jump outside of the already shrunken search space and the knowledge of the current converged search space would be lost. Hence, we calculate opposite particles by using dynamically updated interval boundaries as follows [15]. where is the th position element of the th particle in the population, is the opposite particle of , and are the minimum and maximum values of the th dimension in current search space, respectively, rand are random numbers within , is the box-constraint of the problem, and is the population size, and , indicates the generations.

##### 3.2. Adaptive Constraint Handling

To handle the constraints in solving constrained optimization problems, this paper employs an adaptive penalty method (APM) which was early considered in [18–20]. It aims to help users avoid manually defining the coefficients of penalty functions. In the APM, each constraint of the candidate solutions is monitored. If a constraint seems to be more difficult to satisfy, then a larger penalty coefficient is added.

For each candidate solution, its th constraint violation is computed as follows: where is the th equality constraint, and is the th inequality constraint.

The fitness value of candidate solution is defined by where is the number of constraints, is a penalty coefficient for each constraint, and is defined by where is the average objective function values in the current swarm, and it is computed as where is the population size.

For the penalty coefficient , it determines the scaled factor of the th constraint. Every generation, the is adaptively adjusted as follows. where is the violation of the th constraint averaged over the current swarm, and it is computed by where is the violation of th particle on the th constraint.

##### 3.3. Boundary Search Strategy

For constrained optimization problems, the solution search space can be divided into two parts: feasible space and infeasible space. In some cases, the global optimum is located at the boundaries of feasible space. It is difficult to find this kind of solutions. Because many algorithms can hardly judge the boundaries of feasible space. To tackle this problem, this paper employs a boundary search strategy as follows.

If the current population contains feasible solutions and infeasible solutions (), then we randomly select one feasible solution and one infeasible solution . Based on and , a new solution is generated by where .

Note that the boundary search strategy works when the current population contains feasible and infeasible solutions. Figure 1 clearly illustrates the boundary search. As seen, if the new solution is infeasible, then replace with . Because is nearer to the boundary than . If is feasible, then replace with . Because is nearer to the boundary than .

##### 3.4. The Proposed Approach

The proposed approach (OBPSO) uses a similar procedure to that of opposition-based differential evolution (ODE) for opposition-based population initialization and dynamic opposition [15]. To handle the constraints, we define a new fitness evaluation function as described in (3.6). Moreover, we also use a boundary search strategy to find the solutions located at the margin of the feasible region. The framework of OBPSO is shown in Algorithm 1, where is the current population, OP is the population after using OBL, is the th particle in , is the th particle in GOP, is the probability of opposition, is the population size, is the dimension size, is the interval boundaries of current population, FEs is the number of fitness evaluations, and MAX_FEs is the maximum number of fitness evaluations.

#### 4. Experimental Verifications

##### 4.1. Test Problems

To verify the performance our proposed approach, we employ a set of 13 benchmark functions from the literature [21, 22]. The main characteristics of these benchmark functions are summarized in Table 1. For specific definitions of these functions, please refer to [23].

##### 4.2. Comparison of OBPSO with Similar PSO Algorithms

In this section, we compare the performance of OBPSO with standard PSO, barebones PSO (BPSO), and OBPSO without boundary search strategy (OBPSO-1). To have a fair comparison, the same settings are used for common parameters. The population size is set to 40. For PSO, The inertia weight is set to 0.72984. The acceleration coefficients and are set to 1.49618. The maximum velocity was set to the half range of the search space for each dimension. For OBPSO-1 and OBPSO, the probability of opposition is set to 0.3. For each test functions, both OBPSO and PSO stop running when the number of iterations reaches to 1,000.

Table 2 presents average results of the four PSO algorithms over 30 runs, where Mean represents the mean best function values. As seen, PSO outperforms BPSO on only one problem . BPSO achieves better results than PSO on 7 problems. Both of them can find the global optimum on 5 problems. It demonstrates that the barebones PSO is better than standard PSO for these problems.

For the comparison of OBPSO with BPSO, both of them obtain the same results on 6 problems. For the rest 7 problems, OBPSO performs better than BPSO. It demonstrates that the opposition-based learning is helpful to improve the quality of solutions.

To verify the effects of the boundary search strategy, we compare the performance of OBPSO with OBPSO-1. For the OBPSO-1 algorithm, it does not use the proposed boundary search strategy. As seen, OBPSO outperforms OBPSO-1 on 6 problems, while they obtain the same results for the rest 7 problems. These results demonstrate the effectiveness of the boundary search strategy.

Figure 2 shows the evolutionary processes on four representative problems. It can be seen that OBPSO converges faster than other 3 PSO algorithms. The OBPSO-1 shows faster convergence rate than PSO and BPSO. This confirms that the opposition-based learning is beneficial for accelerating the evolution [15].

##### 4.3. Comparison of OBPSO with Other State-of-the-Art PSO Variants

In this section, we compare the performance of OBPSO with three other PSO variants on the test suite. The involved algorithms and parameter settings are listed as follows.(i)New vector PSO (NVPSO) [4].(ii)Dynamic-objective PSO (RVPSO) [21].(iii)Self-adaptive velocity PSO (SAVPSO) [22].(iv)Our approach OBPSO.

The parameter settings of NVPSO are described in [4]. For RVPSO and SAVPSO, their parameter settings are given in [22]. For OBPSO, we use the same parameter values as described in the previous section. For each test functions, all algorithms stop running when the number of iterations reaches to the maximum value 1,000.

Table 3 presents average results of NVPSO, RVPSO, SAVPSO, and OBPSO over 30 runs, where Mean represents the mean best function values. The comparison results among OBPSO and other algorithms are summarized as in the last row of the table, which means that OBPSO wins in functions, ties in functions, and loses in functions, compared with its competitors.

From the results of Table 3, OBPSO outperforms NVPSO on 7 problems, while NVPSO only achieves better results on a single problem. For the rest 5 problems, both OBPSO and NVPSO can find the global optimum. OBPSO performs better than RVPSO on 8 problems, while both of them obtain the same results for the rest 5 problems. For the comparison of SAVPSO and OBPSO, both of them achieve the same results on 6 problems. For the rest 7 problems, OBPSO wins 6, while SAVPSO wins only 1.

#### 5. Conclusion

This paper proposes a modified barebones particle swarm optimization to solve constrained nonlinear optimization problems. The proposed approach is called OBPSO which employs two novel strategies including opposition-based learning and boundary search. Compared to other improved PSO variants, OBPSO is almost a parameter-free algorithm (except for the probability of opposition). Moreover, an adaptive penalty method is used to handle constraints. Experimental verifications on a set of constrained benchmark functions show that OBPSO achieves a promising performance compared to four other PSO variants. The parameter may affect the performance of OBPSO. To determine the best choice of , different values of will be investigated. This will be conducted in the future work.

#### Acknowledgments

The authors would like to thank the editor and anonymous reviewers for their detailed and constructive comments that helped them to increase the quality of this work. This work is supported by the Science and Technology Plan Projects of Jiangxi Provincial Education Department (nos. GJJ12641, GJJ12633, and GJJ12307), and the National Natural Science Foundation of China (nos. 61070008, 61165004).

#### References

- J. Kennedy and R. Eberhart, “Particle swarm optimization,” in
*Proceedings of the IEEE International Conference on Neural Networks. Part 1*, pp. 1942–1948, December 1995. View at Scopus - K. Meng, H. G. Wang, Z. Y. Dong, and K. P. Wong, “Quantum-inspired particle swarm optimization for valve-point economic load dispatch,”
*IEEE Transactions on Power Systems*, vol. 25, no. 1, pp. 215–222, 2010. View at Publisher · View at Google Scholar · View at Scopus - L. D. S. Coelho, “Gaussian quantum-behaved particle swarm optimization approaches for constrained engineering design problems,”
*Expert Systems with Applications*, vol. 37, no. 2, pp. 1676–1683, 2010. View at Publisher · View at Google Scholar · View at Scopus - C. L. Sun, J. C. Zeng, and J. S. Pan, “An new vector particle swarm optimization for constrained optimization problems,” in
*Proceedings of the International Joint Conference on Computational Sciences and Optimization (CSO '09)*, pp. 485–488, April 2009. View at Publisher · View at Google Scholar · View at Scopus - H. Liu, Z. Cai, and Y. Wang, “Hybridizing particle swarm optimization with differential evolution for constrained numerical and engineering optimization,”
*Applied Soft Computing Journal*, vol. 10, no. 2, pp. 629–640, 2010. View at Publisher · View at Google Scholar · View at Scopus - G. Venter and R. T. Haftka, “Constrained particle swarm optimization using a bi-objective formulation,”
*Structural and Multidisciplinary Optimization*, vol. 40, no. 1–6, pp. 65–76, 2010. View at Publisher · View at Google Scholar - H. Lu and X. Chen, “A new particle swarm optimization with a dynamic inertia weight for solving constrained optimization problems,”
*Information Technology Journal*, vol. 10, no. 8, pp. 1536–1544, 2011. View at Publisher · View at Google Scholar · View at Scopus - M. Daneshyari and G. G. Yen, “Constrained multiple-swarm particle swarm optimization within a cultural framework,”
*IEEE Transactions on Systems, Man, and Cybernetics, Part A*, vol. 18, pp. 1–16, 2011. View at Google Scholar - J. Kennedy, “Bare bones particle swarms,” in
*Proceedings of the IEEE Swarm Intelligence Symposium (SIS '03)*, pp. 80–87, 2003. - M. G. H. Omran, A. P. Engelbrecht, and A. Salman, “Bare bones differential evolution,”
*European Journal of Operational Research*, vol. 196, no. 1, pp. 128–139, 2009. View at Publisher · View at Google Scholar - R. A. Krohling and E. Mendel, “Bare bones particle swarm optimization with Gaussian or cauchy jumps,” in
*Proceedings of the IEEE Congress on Evolutionary Computation (CEC '09)*, pp. 3285–3291, May 2009. View at Publisher · View at Google Scholar · View at Scopus - H. R. Tizhoosh, “Opposition-based learning: a new scheme for machine intelligence,” in
*Proceedings of the International Conference on Computational Intelligence for Modelling, Control and Automation (CIMCA '05) and International Conference on Intelligent Agents, Web Technologies and Internet Commerce (IAWTIC '05)*, pp. 695–701, November 2005. View at Scopus - Y. Shi and R. C. Eberhart, “A modified particle swarm optimizer,” in
*Proceedings of the IEEE International Conference on Evolutionary Computation (ICEC '98)*, pp. 69–73, May 1998. View at Scopus - A. P. Engelbrecht, “Heterogeneous particle swarm optimization,” in
*Proceedings of the International Conference on Swarm Intelligence*, pp. 191–202, 2010. - R. S. Rahnamayan, H. R. Tizhoosh, and M. M. A. Salama, “Opposition-based differential evolution,”
*IEEE Transactions on Evolutionary Computation*, vol. 12, no. 1, pp. 64–79, 2008. View at Publisher · View at Google Scholar · View at Scopus - H. Wang, Z. Wu, S. Rahnamayan, Y. Liu, and M. Ventresca, “Enhancing particle swarm optimization using generalized opposition-based learning,”
*Information Sciences*, vol. 181, no. 20, pp. 4699–4714, 2011. View at Publisher · View at Google Scholar - H. Wang, Z. Wu, and S. Rahnamayan, “Enhanced opposition-based differential evolution for solving high-dimensional continuous optimization problems,”
*Soft Computing*, vol. 15, no. 11, pp. 2127–2140, 2011. View at Publisher · View at Google Scholar · View at Scopus - A. C. C. Lemonge and H. J. C. Barbosa, “An adaptive penalty scheme for genetic algorithms in structural optimization,”
*International Journal for Numerical Methods in Engineering*, vol. 59, no. 5, pp. 703–736, 2004. View at Publisher · View at Google Scholar · View at Scopus - E. K. da Silva, H. J. C. Barbosa, and A. C. C. Lemonge, “An adaptive constraint handling technique for differential evolution with dynamic use of variants in engineering optimization,”
*Optimization and Engineering*, vol. 12, no. 1-2, pp. 31–54, 2011. View at Publisher · View at Google Scholar - X. Pan, Y. Cao, and Q. Pu, “Improved particle swarm optimization with adaptive constraint handling for engineering optimization,”
*Journal of Information and Computational Science*, vol. 8, no. 15, pp. 3507–3514, 2011. View at Google Scholar · View at Scopus - H. Y. Lu and W. Q. Chen, “Dynamic-objective particle swarm optimization for constrained optimization problems,”
*Journal of Combinatorial Optimization*, vol. 12, no. 4, pp. 409–419, 2006. View at Publisher · View at Google Scholar - H. Y. Lu and W. Q. Chen, “Self-adaptive velocity particle swarm optimization for solving constrained optimization problems,”
*Journal of Global Optimization*, vol. 41, no. 3, pp. 427–445, 2008. View at Publisher · View at Google Scholar - J. J. Liang, T. P. Runarsson, E. Mezura-Montes et al., “Problem definitions and evaluation criteria for the CEC 2006, special session on constrained real-parameter optimization,” Tech. Rep., Nanyang Technological University, Singapore, 2006. View at Google Scholar