- About this Journal ·
- Abstracting and Indexing ·
- Aims and Scope ·
- Annual Issues ·
- Article Processing Charges ·
- Articles in Press ·
- Author Guidelines ·
- Bibliographic Information ·
- Citations to this Journal ·
- Contact Information ·
- Editorial Board ·
- Editorial Workflow ·
- Free eTOC Alerts ·
- Publication Ethics ·
- Reviewers Acknowledgment ·
- Submit a Manuscript ·
- Subscription Information ·
- Table of Contents

The Scientific World Journal

Volume 2013 (2013), Article ID 409167, 8 pages

http://dx.doi.org/10.1155/2013/409167

## A New Logistic Dynamic Particle Swarm Optimization Algorithm Based on Random Topology

^{1}School of Computer Science and Engineering, Southeast University, Nanjing 211189, China^{2}Provincial Key Laboratory for Computer Information Processing Technology, Soochow University, Suzhou 215006, China

Received 17 April 2013; Accepted 19 May 2013

Academic Editors: P. Agarwal, S. Balochian, and V. Bhatnagar

Copyright © 2013 Qingjian Ni and Jianming Deng. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

#### Abstract

Population topology of particle swarm optimization (PSO) will directly affect the dissemination of optimal information during the evolutionary process and will have a significant impact on the performance of PSO. Classic static population topologies are usually used in PSO, such as fully connected topology, ring topology, star topology, and square topology. In this paper, the performance of PSO with the proposed random topologies is analyzed, and the relationship between population topology and the performance of PSO is also explored from the perspective of graph theory characteristics in population topologies. Further, in a relatively new PSO variant which named logistic dynamic particle optimization, an extensive simulation study is presented to discuss the effectiveness of the random topology and the design strategies of population topology. Finally, the experimental data are analyzed and discussed. And about the design and use of population topology on PSO, some useful conclusions are proposed which can provide a basis for further discussion and research.

#### 1. Introduction

Particle swarm optimization (briefed as PSO) is a kind of bionic evolutionary algorithm which rooted in imitation of behavioral mechanisms in populations such as birds and fish stocks and has been widely used in engineering field as optimization method [1–3].

In the PSO algorithms, the particles evolve according to their own experience and the experience of the neighborhood particles. During the evolutionary process, particles identify their own neighborhood according to the population topology and then learn from each other and update the positions of the particles. Therefore, population topology determines the form of information sharing among particles and thus has a very important impact on the solving performance of PSO algorithms. Therefore, it is important to explore the population topologies of PSO algorithms. This will produce a deep understanding of the working mechanism of PSO algorithms and thus improve the solving performance.

In the PSO algorithms, the most common used static population topologies are the fully connected topology (Gbest model) and the ring topology (lbest model) which are also first proposed [4]. Since then, researchers have proposed different population topologies in succession. Kennedy carried out a preliminary analysis of four static population topologies [5]. Suganthan adjusted the neighborhood structure of particles through calculating distances between particles in the evolutionary process [6]. Mendes et al. detailly analyzed the relationship between the population topology and a class of PSO variant [7–9]. Clerc initially attempted to adopt the random topology [10]. However, these studies concerned population topologies paid more attention to the static classic population topology, and research of random population topologies is relatively small. As the population topology directly affect the exchange of information between the particles, it is necessary to design the suitable population topology according to the characteristics of different types of applications. Therefore, it is necessary to explore the population topologies in depth from a theoretical and experimental point of view.

In this paper, in a relatively new PSO variant which named logistic dynamic particle optimization, we analyze the linkages between population topologies and the performance of PSO algorithms from graph theory and experimental point of view. The rest of the paper is organized as follows. In Section 2, we describes the PSO variant which are used in the paper. Section 3 describes the classic population topologies and introduces the proposed random population topologies. In Section 4, we have presented the experimental analysis and comparative performance between the classic and the proposed random population topologies. Section 5 concludes the paper.

#### 2. The PSO Variants

PSO is a population-based method which is similar to other evolutionary computation methods. The individual in the PSO population is called the particle, and particles generally have the speed and position in most PSO variants.

##### 2.1. The Canonical PSO

Based on the earlier version of PSO, Clerc developed the PSO with the compression factor [11]. This PSO variants have been widely used in practical applications, and the velocity and position update of particles in this variant are as follows:

In (1), is the th dimensional component of particle s velocity attribute, and are two positive acceleration factors, and are random number generating functions between 0 and 1, is the th dimensional component of the particle s position property, is the th dimensional component of the best position that particle obtained, and is the th dimensional component of the best position that the whole population obtained. In actual use, usually in (2) is set to 0.729, and is often set to 4.1.

The right part of the equation1 can be understood as particle’s memory, cognitive and social cognition. Velocity of particles is precisely through this three-part interaction effects thereby, position of particles is updated.

##### 2.2. The Dynamic Probabilistic Particle Swarm Optimization

In the previous PSO variants, particles usually have both velocity attribute and position attribute. Kennedy first proposed a new PSO variant without velocity attribute, which named Gaussian dynamic particle swarm optimization [12]. Ni and Deng carried out a further study on the PSO variants without velocity [13]. This type of algorithm variants can be called dynamic probabilistic particle swarm optimization (briefed as DPPSO). In the DPPSO algorithms, particles have no velocity attributes, and the update of particles’ positions is reorganized as follows:

In (4), (5), and (6), represents the current evolution generation of particle, is the index number of particle, is the index number of particle’s neighborhood, represents the number of particle’s neighborhood particles, is the optimum position of the particle’s neighborhood that numbered , and is the number of the particles s dimension. is an abbreviation of centralized tendency, which is a -dimensional vector and is determined by the particle’s current location and neighborhood particles’ optimal positions. is an abbreviation of outlier trend, which is also a -dimensional vector and is determined by the particle’s current location and neighborhood particles’ optimal positions. In (4), , , and are generally preferable to positive constants. is a dynamic probabilistic evolutionary operator which is the random number generator function which satisfies a specific distribution, and this particular distribution may be provided by a Gaussian distribution or a logistic distribution and so forth.

In (4), the calculation of particle’s position in new generation is decided jointly by the right four parts. The first part is the memory of particles on the self-position. The second part is the trend of the particles along the previous direction of movement of the “flying.” The third part means the influence of neighborhood particles’ experience to the new generation position, this part of the calculations needs neighborhood particles’ experience, and this part determines the degree of influence of the neighborhood particles’ experience. The part IV reflects the impact of differences in best position between particles on the next generation position.

The performance of DPPSO variants is different when using different dynamic probabilistic evolutionary operator [13]. DPPSO-Gaussian (briefed as GDPS) has faster convergence speed in the early evolution. DPPSO-Cauchy may get better solutions on certain issues, but the performance is unstable. DPPSO-Logistic (logistic dynamic particle swarm optimization, briefed as LDPSO) still shows good exploration ability in the later evolution. For DPPSO variants, the calculation of and will use the experience of each neighborhood particles which can also be seen from (5) and (6). The optimal information of neighborhood particles could be fully utilized, so the research of population topology in DPPSO variants is more important.

#### 3. Population Topologies

##### 3.1. The Classic Population Topologies

Figure 1 shows the fully connected topology, ring topology, and star topology. Fully connected topology and ring topology are two commonly used topologies which are also called Gbest model and Lbest model. In the fully connected topology, particle’s neighborhood contains all particles in the population. And in the evolutionary process, only the particle that obtains the optimal position is considered in the entire population. PSO algorithms with this topology converge very fast but easy to fall into local optimum.

For a ring topology, typically particle’s neighborhood includes the particles on both sides of one or a few particles. In this topology, the exchange of information is relatively slow within the population, but once a particle searched for an optimal location, the information eventually will slowly spread to the entire population.

For a star topology, one particle is connected with all the other particles, and other particles only connect with the particle. In addition to a central particle, other particles are independent of each other, the dissemination of information must be passed through the central particle.

However, the information dissemination mechanism of the social groups is not static throughout the whole evolutionary process, and tends to have a certain degree of randomness and dynamic characteristics. Mendes studied random population topology and confirmed that population topology directly affects the execution performance of PSO algorithms [9].

##### 3.2. Graph Theory Characteristics of Population Topology

The population topology of the PSO algorithm can be abstracted into a connected undirected graph, represented by the symbol , where is the set of vertices, is the set of edges, and the number of vertices is denoted by . For any two points and in , denotes the distance from to , that is, the length of the shortest path between two points.

*Definition 1 (average degree). *The degree of the vertex is the number of its adjacent vertices, denoted by . Average degree of undirected connected graph is calculated by

*Definition 2 (average clustering coefficient). *The local clustering coefficient of vertex , equals to the number of edges which can be connected between vertices associated with vertex , divided by the maximum number of edges between these vertices. The average clustering coefficient of a graph is the arithmetic mean of the local clustering coefficient of all vertices, which can be calculated by

The average degree of population topology means the average number of particles’ neighborhood particles; it represents the degree of socialization of population. A small number of neighborhood particles means that the particle is both difficult to obtain information from the population and difficult to influence other particles. On the contrary, a particle which have large number of neighborhood particles can get a lot of information available in the population, and such a particle has a greater influence in the population. In the common used population topologies, fully connected topology has the largest average degree which is equal to the population size minus 1. The ring topology has the minimum average degree; the average degree of a ring topology such as in Figure 1 is equal to 2. The local clustering coefficient is the ratio of the number of connections between the actual existence and the possible existence, and the average clustering coefficient represents the degree of aggregation of the vertices in a graph which is the average of the local clustering coefficient of all vertices. In this paper, we will analyze the role of the population topology based on the previous graph theory characteristics.

##### 3.3. Random Population Topologies

Clerc initially attempted to proposed a method of random population topology [10]. The basic idea is to generate a random topology by selecting the neighborhood particles randomly for each particle. The concrete steps could be described as Algorithm 1.

In the resulting matrix of Algorithm 1, means that the particles and are connected. By this method, a random topology could be generated with an average degree slightly larger than . In the random topology by this method, the distribution of degree is the sum of independent Bernoulli random variables which is described in [10]

In this paper, in order to ensure the connectivity of undirected graph which is corresponding to the generating random topology, we proposed a new generating method of random topology based on Clerc’s method. The basic idea is: in selecting the neighborhood particles for each particle if the selection is itself, reselect in order to reduce the probability of particles isolated; after the random population is generated, use the Dijkstra algorithm to compute the distance between the particles if there exist the unconnected particles in the generated topology, then add an edge between these unconnected particles, and retest. The improved random population topology generating method is as Algorithm 2.

#### 4. Experiment and Analysis

##### 4.1. Experiment Setting

Two sets of experiments were conducted which used the canonical PSO (briefed as CPSO) and the DPPSO-Logistic, respectively, that are described in Section 2. For CPSO, set , . For DPPSO-Logistic, set , , and .

The algorithms were used to solve five benchmark functions, which is defined in Table 1. These functions consist of Ackley, Schwefel, Schaffer’s F6, Rastrigin, and Sphere. Table 2 shows the settings of these functions.

In the experiment, the population size is set to 20, in addition to the the Schaffer’s F6 function of the dimension 2, the remaining functions are carried out in the case of 30-dimensional test, and the frequency of repeated experiments is 50.

The performance of the algorithms will be evaluated by the following aspects:(i)in the case of a certain number of iterations, compare the accuracy (briefed as Perform.) of the optimal fitness value in each case. These values reflect the quality of the optimal solution obtained in the last; (ii)in the case of a certain number of iterations, compare the success rates (briefed as Prop.) which means that the algorithms achieve the accuracy (accepted error) that is defined in Table 2. These data reflect the stability of the algorithms; (iii)in the case of a certain number of iterations, compare the evolutionary trends of various algorithms, these figures reflect the evolution of the optimal solution in the evolutionary process.

##### 4.2. Comparison between Random Topologies and Different Average Degrees

For the random topology, the first set of experiments used the canonical PSO algorithm to solve the five benchmark functions. By changing the value, we generated random topologies with different average degrees for comparison and evaluation. Experimental results are shown in Table 3.

As can be seen from Table 3, with the increase in the value of , the indicators of Perform. and Prop. have improved. For multimodal functions such as Schwefel, Schaffer’s F6, Rastrigin, and Ackley function, with the increase in the value of , when the value is from 3 to 4 (i.e., the average degree of the random topology is between 5 and 7, substantially in about 6), the PSO algorithm could get better performance. And when the value is larger, the performance is usually poor. For multimodal functions, when a particle has found a local optimal solution, if the interaction between particles is more, the dissemination of information will be very quickly, and the entire population is susceptible to rapid convergence to the local optimal solution. Therefore, too large average degree of population topology is not conducive to find the global optimal solution for a multimodal function.

Taking these factors together, when the value of is at about 4, the PSO algorithm has a more satisfactory performance for most benchmark functions. Accordingly, the following experiments will generate random topologies which the value is 4, and compare these topologies with other classic static population topologies.

##### 4.3. Comparison between Random Topologies and Classic Topologies

The second set of experiments used the DPPSO-Logistic algorithm to solve the five benchmark functions on the fully connected topology, ring topology, star topology, and random topology, wherein value is set to 4 to generate random topologies. Comparison and evaluation are conducted by the evolutionary trends of algorithms with various population topologies. In the figures of evolutionary trends, different line types expressed different DPPSO-Logistic algorithms with various population topologies.

For Sphere function (Figure 2), the performance of ring topology and star topology is poor, and the fully connected topology and random topology show better search ability. For Schaffer’s F6 function (Figure 2), random topology is significantly better than the three classic neighborhood topologies.

For Rastrigin function (Figure 3), random topology is superior to the three classic population topologies; in the early stages of evolution, the convergence speed of the random topology is almost the same as the fully connected topology; however, in the later stage of evolution, random topology shows a larger advantage, and the end result is better than the fully connected topology.

For Schwefel function (Figure 3), the performance of the ring topology is poor; the fully connected topology convergence fast in the early stage of evolution, but the end result is poor; the star topology has achieved good results; the convergence speed of random topology is second only to the fully connected topology in the early evolution stage, and the final result of random topology is better than the other classic topologies.

For Ackley function (Figure 4), the fully connected topology and the star topology show poor performance; the ring topology performs better; the random topology shows obvious advantage in both the convergence speed and the final result.

Overall, according to the convergence speed, random topology is relatively stable in the early stages of evolution, faster in the midstages of evolution, and shows a distinct advantage in the late stages of evolution. From the view of final result, the PSO algorithms using random topology demonstrate remarkable performance.

For unimodal function (such as Sphere function), because there is no problem of falling into a local optimum, the close ties between particles can make faster convergence and achieve better results. Therefore, the performance of the fully connected topology and random topology with a relatively high average degree is ideal. In the case of random topology for unimodal function, the convergence speed and the solution will be better with the greater average degree of population topology.

For multimodal function, it can be seen that the convergence speed will be faster when the average degree of population topology is increasing. If the average degree of population topology is too high, it is easy to fall into local optimum. On the average degree after 6, the optimal solution quality of most algorithms begins to decrease. When the average degree is between 5 and 7 ( is set to 4), the performance of algorithms is usually ideal.

#### 5. Conclusion

In this paper, we propose an improved method of generating random topology based on previous research. And we carry out in-depth research of the performance of the algorithms using random topology based on the canonical PSO and DPPSO-Logistic, respectively. Combined with experimental results, we conduct the analysis and interpretation of the performance of the algorithms from the perspective of graph theory. And empirical laws in generating random topologies are given according to our experimental results and theoretical analysis.

On the whole, relative to the three classic population topologies (fully connected topology, ring topology, and star topology), the algorithm has obvious advantages which is using the proposed random topology. Further work will include the theoretical analysis of different population topologies, as well as dynamic population topology strategy which is designed in accordance with the conclusions of this paper.

#### Acknowledgments

This paper is supported by Provincial Key Laboratory for Computer Information Processing Technology, Soochow University, Suzhou, China, and NSFC (Grant no. 61170164).

#### References

- M. R. AlRashidi and M. E. El-Hawary, “A survey of particle swarm optimization applications in electric power systems,”
*IEEE Transactions on Evolutionary Computation*, vol. 13, no. 4, pp. 913–918, 2009. View at Publisher · View at Google Scholar · View at Scopus - A. A. Esmin, R. A. Coelho, and S. Matwin, “A review on particle swarm optimization algorithm and its variants to clustering highdimensional data,”
*Artificial Intelligence Review*, 2013. View at Publisher · View at Google Scholar - R. V. Kulkarni and G. K. Venayagamoorthy, “Particle swarm optimization in wireless-sensor networks: a brief survey,”
*IEEE Transactions on Systems, Man and Cybernetics C*, vol. 41, no. 2, pp. 262–267, 2011. View at Publisher · View at Google Scholar · View at Scopus - J. Kennedy and R. Eberhart, “Particle swarm optimization,” in
*Proceedings of the IEEE International Conference on Neural Networks*, vol. 4, pp. 1942–1948, December 1995. View at Scopus - J. Kennedy, “Small worlds and mega-minds: effects of neighborhood topology on particle swarm performance,” in
*Proceedings of the Congress on Evolutionary Computation (CEC '99)*, vol. 3, IEEE, 1999. - P. N. Suganthan, “Particle swarm optimiser with neighbourhood operator,” in
*Proceedings of the Congress on Evolutionary Computation (CEC '99)*, vol. 3, IEEE, 1999. - R. Mendes, J. Kennedy, and J. Neves, “The fully informed particle swarm: simpler, maybe better,”
*IEEE Transactions on Evolutionary Computation*, vol. 8, no. 3, pp. 204–210, 2004. View at Publisher · View at Google Scholar · View at Scopus - M. A. M. de Oca and T. Stützle, “Convergence behavior of the fully informed particle swarm optimization algorithm,” in
*Proceedings of the 10th Annual Genetic and Evolutionary Computation Conference (GECCO '08)*, pp. 71–78, ACM, July 2008. View at Scopus - R. Mendes,
*Population topologies and their influence in particle swarm performance [Ph.D. dissertation]*, Universidade do Minho, 2004. - M. Clerc, “Back to random topology,” Tech. Rep. 2007, 2007.
- M. Clerc and J. Kennedy, “The particle swarm-explosion, stability, and convergence in a multidimensional complex space,”
*IEEE Transactions on Evolutionary Computation*, vol. 6, no. 1, pp. 58–73, 2002. View at Publisher · View at Google Scholar · View at Scopus - J. Kennedy, “Dynamic-probabilistic particle swarms,” in
*Proceedings of the Conference on Genetic and Evolutionary Computation Conference*, pp. 201–207, ACM, June 2005. View at Publisher · View at Google Scholar · View at Scopus - Q. Ni and J. Deng, “Two improvement strategies for logistic dynamic particle swarm optimization,” in
*Adaptive and Natural Computing Algorithms*, pp. 320–329, Springer, 2011. View at Publisher · View at Google Scholar · View at Scopus