- About this Journal ·
- Abstracting and Indexing ·
- Aims and Scope ·
- Annual Issues ·
- Article Processing Charges ·
- Articles in Press ·
- Author Guidelines ·
- Bibliographic Information ·
- Citations to this Journal ·
- Contact Information ·
- Editorial Board ·
- Editorial Workflow ·
- Free eTOC Alerts ·
- Publication Ethics ·
- Reviewers Acknowledgment ·
- Submit a Manuscript ·
- Subscription Information ·
- Table of Contents
The Scientific World Journal
Volume 2014 (2014), Article ID 136920, 17 pages
Bare-Bones Teaching-Learning-Based Optimization
1School of Computer Science and Engineering, Xi’an University of Technology, Xi’an 710048, China
2School of Physics and Electronic Information, Huaibei Normal University, Huaibei 235000, China
Received 20 February 2014; Accepted 7 April 2014; Published 10 June 2014
Academic Editors: S. Balochian and Y. Zhang
Copyright © 2014 Feng Zou et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Teaching-learning-based optimization (TLBO) algorithm which simulates the teaching-learning process of the class room is one of the recently proposed swarm intelligent (SI) algorithms. In this paper, a new TLBO variant called bare-bones teaching-learning-based optimization (BBTLBO) is presented to solve the global optimization problems. In this method, each learner of teacher phase employs an interactive learning strategy, which is the hybridization of the learning strategy of teacher phase in the standard TLBO and Gaussian sampling learning based on neighborhood search, and each learner of learner phase employs the learning strategy of learner phase in the standard TLBO or the new neighborhood search strategy. To verify the performance of our approaches, 20 benchmark functions and two real-world problems are utilized. Conducted experiments can been observed that the BBTLBO performs significantly better than, or at least comparable to, TLBO and some existing bare-bones algorithms. The results indicate that the proposed algorithm is competitive to some other optimization algorithms.
Many real-life optimization problems are becoming more and more complex and difficult with the development of scientific technology. So how to resolve these complex problems in an exact manner within a reasonable time cost is very important. The traditional optimization algorithms are difficult to solve these complex nonlinear problems. In recent years, nature-inspired optimization algorithms which simulate natural phenomena and have different design philosophies and characteristics, such as evolutionary algorithms [1–3] and swarm intelligence algorithms [4–7], are a research field which simulates different natural phenomena to solve a wide range of problems. In these algorithms the convergence rate of the algorithm is given prime importance for solving real-world optimization problems. The ability of the algorithms to obtain the global optima value is one aspect and the faster convergence is the other aspect.
As a stochastic search scheme, TLBO [8, 9] is a newly population-based algorithm based on swarm intelligence and has characters of simple computation and rapid convergence; it has been extended to the function optimization, engineering optimization, multiobjective optimization, clustering, and so forth [9–17]. TLBO is a parameter-free evolutionary technique and is also gaining popularity due to its ability to achieve better results in comparatively faster convergence time to genetic algorithms (GA) , particle swarm optimizer (PSO) , and artificial bee colony algorithm (ABC) . However, in evolutionary computation research there have been always attempts to improve any given findings further and further. This work is an attempt to improve the convergence characteristics of TLBO further without sacrificing the accuracies obtained in TLBO and in some occasions trying to even better the accuracies. The aims of this paper are of threefold. First, authors propose an improved version of TLBO, namely, BBTLBO. Next, the proposed technique is validated on unimodal and multimodal functions based on different performance indicators. The result of BBTLBO is compared with other algorithms. Results of both the algorithms are also compared using statistical paired -test. Thirdly, it is applied to solve the real-world optimization problem.
The remainder of this paper is organized as follows. The TLBO algorithm is introduced in Section 2. Section 3 presents a brief overview of some recently proposed bare-bones algorithms. Section 4 describes the improved teaching-learning-based optimization algorithm using neighborhood search (BBTLBO). Section 5 presents the tests on several benchmark functions and the experiments are conducted along with statistical tests. The applications for training artificial neural network are shown in Section 6. Conclusions are given in Section 7.
2. Teaching-Learning-Based Optimization
Rao et al. [8, 9] first proposed a novel teaching-learning-based optimization (TLBO) inspired from the philosophy of teaching and learning. The TLBO algorithm is based on the effect of the influence of a teacher on the output of learners in a class which is considered in terms of results or grades. The process of working of TLBO is divided into two parts. The first part consists of “teacher phase” and the second part consists of “learner phase.” The “teacher phase” means learning from the teacher and the “learner phase” means learning through the interaction between learners.
A good teacher is one who brings his or her learners up to his or her level in terms of knowledge. But in practice this is not possible and a teacher can only move the mean of a class up to some extent depending on the capability of the class. This follows a random process depending on many factors. Let be the mean and let be the teacher at any iteration. will try to move mean toward its own level, so now the new mean will be designated as . The solution is updated according to the difference between the existing and the new mean according to the following expression: where TF is a teaching factor that decides the value of mean to be changed and is a random vector in which each element is a random number in the range . The value of TF can be either 1 or 2, which is again a heuristic step and decided randomly with equal probability as
Learners increase their knowledge by two different means: one through input from the teacher and the other through interaction between themselves. A learner interacts randomly with other learners with the help of group discussions, presentations, formal communications, and so forth. A learner learns something new if the other learner has more knowledge than him or her. Learner modification is expressed as
As explained above, the pseudocode for the implementation of TLBO is summarized in Algorithm 1.
3. Bare-Bones Algorithm
In this section, we only presented a brief overview of some recently proposed bare-bones algorithms.
3.1. BBPSO and BBExp
PSO is a swarm intelligence-based algorithm, which is inspired by the behavior of birds flocking . In PSO, each particle is attracted by its personal best position and the global best position found so far. Theoretical studies [18, 19] proved that each particle converges to the weighted average of and : where and are two leaning factors in PSO.
Based on the convergence characteristic of PSO, Kennedy  proposed a new PSO variant called bare-bones PSO (BBPSO). Bare-bones PSO retains the standard PSO social communication but replaces dynamical particle update with sampling from a probability distribution based on and as follows: where is the th dimension of the th particle in the population and represents a Gaussian distribution with mean and standard deviation .
Kennedy  proposed also an alternative version of the BBPSO, denoted by BBExp, where (5) is replaced by where is a random value within for the th dimension. For the alternative mechanism, there is a 50% chance that the search process is focusing on the previous best positions.
3.2. BBDE, GBDE, and MGBDE
Inspired by the BBPSO and DE, Omran et al.  proposed a new and efficient DE variant, called bare-bones differential evolution (BBDE). The BBDE is a new, almost parameter-free optimization algorithm that is a hybrid of the bare-bones particle swarm optimizer and differential evolution. Differential evolution is used to mutate, for each particle, the attractor associated with that particle, defined as a weighted average of its personal and neighborhood best positions. For the BBDE, the individual is updated as follows: where , , and are three indices chosen from the set with , rand (0, 1) is a random value within for the th dimension, and is defined by where and are personal best position and the global best position, , is a random value within for the th dimension.
Based on the idea that the Gaussian sampling is a fine tuning procedure which starts during exploration and is continued to exploitation, Wang et al.  proposed a new parameter-free DE algorithm, called GBDE. In the GBDE, the mutation strategy uses a Gaussian sampling method which is defined by where represents a Gaussian distribution with mean and standard deviation and CR is the probability of crossover.
To balance the global search ability and convergence rate, Wang et al.  proposed a modified GBDE (called MGBDE). The mutation strategy uses a hybridization of GBDE and DE/best/1 as follows:
4. Proposed Algorithm: BBTLBO
The bare-bones PSO utilizes this information by sampling candidate solutions, normally distributed around the formally derived attractor point. That is, the new position is generated by a Gaussian distribution for sampling the search space based on the and the at the current iteration. As a result, the new position will be centered around the weighted average of and . Generally speaking, at the initial evolutionary stages, the search process focuses on exploration due to the large deviation. With an increasing number of generations, the deviation becomes smaller, and the search process will focus on exploitation. From the search behavior of BBPSO, the Gaussian sampling is a fine tuning procedure which starts during exploration and is continued to exploitation. This can be beneficial for the search of many evolutionary optimization algorithms. Additionally, the bare-bones PSO has no parameters to be tuned.
Based on a previous explanation, a new bare-bones TLBO (BBTLBO) with neighborhood search is proposed in this paper. In fact, for TLBO, if the new learner has a better function value than that of the old learner, it is replaced with the old one in the memory. Otherwise, the old one is retained in the memory. In other words, a greedy selection mechanism is employed as the selection operation between the old and the candidate one. Hence, the new teacher and the new learner are the global best and learner’s personal best found so far, respectively. The complete flowchart of the BBTLBO algorithm is shown in Figure 1.
4.1. Neighborhood Search
It is known that birds of a feather flock together and people of a mind fall into the same group. Just like evolutionary algorithms themselves, the notion of neighborhood is inspired by nature. Neighborhood technique is an efficient method to maintain diversity of the solutions. It plays an important role in evolutionary algorithms and is often introduced by researchers in order to allow maintenance of a population of diverse individuals and improve the exploration capability of population-based heuristic algorithms [23–26]. In fact, learners with similar interests form different learning groups. Because of his or her favor characteristic, the learner maybe learns from the excellent individual in the learning group.
For the implementation of grouping, various types of connected distances may be used. Here we have used a ring topology  based on the indexes of learners for the sake of simplicity. In a ring topology, the first individual is the neighbor of the last individual and vice versa. Based on the ring topology, a -neighborhood radius is defined, where is a predefined integer number. For each individual, its -neighborhood radius consists of individuals (including oneself), which are . That is, the neighborhood size is for a -neighborhood. For simplicity, is set to 1 (Figure 2) in our algorithm. This means that there are 3 individuals in each learning group. Once groups are constructed, we can utilize them for updating the learners of the corresponding group.
4.2. Teacher Phase
To balance the global and local search ability, a modified interactive learning strategy is proposed in teacher phase. In this learning phase, each learner employs an interactive learning strategy (the hybridization of the learning strategy of teacher phase in the standard TLBO and Gaussian sampling learning) based on neighborhood search.
In BBTLBO, the updating formula of the learning for a learner in teacher phase is proposed by the hybridization of the learning strategy of teacher phase and the Gaussian sampling learning as follows: where called the hybridization factor is a random number in the range for the th dimension, and are the existing neighborhood best solution and the neighborhood mean solution of each learner, and TF is a teaching factor which can be either 1 or 2 randomly.
In the BBTLBO, there is a chance that the th dimension of the th learner in the population follows the behavior of the learning strategy of teacher phase, while the remaining follow the search behavior of the Gaussian sampling in teacher phase. This will be helpful to balance the advantages of fast convergence rate (the attraction of the learning strategy of teacher phase) and exploration (the Gaussian sampling) in BBTLBO.
4.3. Learner Phase
At the same time, in the learner phase, a learner interacts randomly with other learners for enhancing his or her knowledge in the class. This learning method can be treated as the global search strategy (shown in (3)).
In this paper, we introduce a new learning strategy in which each learner learns from the neighborhood teacher and the other learner selected randomly of his or her corresponding neighborhood in learner phase. This learning method can be treated as the neighborhood search strategy. Let represent the interactive learning result of the learner . This neighborhood search strategy can be expressed as follows: where and are random vectors in which each element is a random number in the range , is the teacher of the learner ’s corresponding neighborhood, and the learner is selected randomly from the learner’s corresponding neighborhood.
In BBTLBO, each learner is probabilistically learning by means of the global search strategy or the neighborhood search strategy in learner phase. That is, about 50% of learners in the population execute the learning strategy of learner phase in the standard TLBO (shown in (3)), while the remaining 50% execute neighborhood search strategy (shown in (12)). This will be helpful to balance the global search and local search in learner phase.
Moreover, compared to the original TLBO, BBTLBO only modifies the learning strategies. Therefore, both the original TLBO and BBTLBO have the same time complexity , where NP is the number of the population, is the number of dimensions, and is the maximum number of generations.
As explained above, the pseudocode for the implementation of BBTLBO is summarized in Algorithm 2.
5. Functions Optimization
In this section, to illustrate the effectiveness of the proposed method, 20 benchmark functions are used to test the efficiency of BBTLBO. To compare the search performance of BBTLBO with some other methods, other different algorithms are also simulated in the paper.
5.1. Benchmark Functions
The details of 20 benchmark functions are shown in Table 1. Among 20 benchmark functions, to are unimodal functions, and to are multimodal functions. The searching range and theory optima for all functions are also shown in Table 1.
5.2. Parameter Settings
All the experiments are carried out on the same machine with a Celoron 2.26 GHz CPU, 2 GB memory, and Windows XP operating system with Matlab 7.9. For the purpose of reducing statistical errors, each algorithm is independently simulated 50 times. For all algorithms, the population size was set to 20. Population-based stochastic algorithms use the same stopping criterion, that is, reaching a certain number of function evaluations (FEs).
5.3. Effect of Variation in Parameter
The hybridization factor u is set to . Comparative tests have been performed using different . In our experiment, the maximal FEs are used as ended condition of algorithm, namely, 40,000 for all test functions. Table 2 shows the mean optimum solutions and the standard deviation of the solutions obtained using different hybridization factor in the 50 independent runs. The best results among the algorithms are shown in bold. Figure 3 presents the representative convergence graphs of different benchmark functions in terms of the mean fitness values achieved by using different hybridization factor on all test functions. Due to the tight space limitation, some sample graphs are illustrated.
The comparisons in Table 2 and Figure 3 show that when the hybridization factor is set to 0.9, BBTLBO offers the best performance on 20 test functions. Hence, the hybridization factor is set to 0.9 in the following experiments.
5.4. Comparison of BBTLBO with Some Similar Bare-Bones Algorithms
In this section, we compare BBTLBO with five other recently proposed three bare-bones DE variants and two bare-bones PSO algorithms. Our experiment includes two series of comparisons in terms of the solution accuracy and the solution convergence (convergence speed and success rate). We compared the performance of BBTLBO with other similar bare-bones algorithms, including BBPSO , BBExp , BBDE , GBDE , and MGBDE .
5.4.1. Comparisons on the Solution Accuracy
In our experiment, the maximal FEs are used as ended condition of algorithm, namely, 40,000 for all test functions. The results are shown in Table 3 in terms of the mean optimum solution and the standard deviation of the solutions obtained in the 50 independent runs by each algorithm on 20 test functions. The best results among the algorithms are shown in bold. Figure 4 presents the convergence graphs of different benchmark functions in terms of the mean fitness values achieved by 7 algorithms for 50 independent runs. Due to the tight space limitation, some sample graphs are illustrated.
From Table 3 it can be observed that the mean optimum solution and the standard deviation of all algorithms perform well for the functions and . Although BBExp performs better than BBTLBO on function and MGBDE performs better than BBTLBO on function , our approach BBTLBO achieves better results than other algorithms on the rest of test functions. Table 3 and Figure 4 conclude that the BBTLBO has a good performance of the solution accuracy for test functions in this paper.
5.4.2. Comparison of the Convergence Speed and SR
In order to compare the convergence speed and successful rate (SR) of different algorithms, we select a threshold value of the objective function for each test function. For other functions, the threshold values are listed in Table 4. In our experiment, the stopping criterion is that each algorithm is terminated when the best fitness value so far is below the predefined threshold value ( Value) or the number of FEs reaches to the maximal FEs 40,000. The results are shown in Table 4 in terms of the mean number of FEs (MFEs) required to converge to the threshold and successful rate (SR) in the 50 independent runs. “NaN” represents that no runs of the corresponding algorithm converged below the predefined threshold before meeting the maximum number of FEs. The best results among the six algorithms are shown in boldface.
From Table 5 it can be observed that all algorithms hardly converge to the threshold for unimodal functions , , , and and multimodal functions , , and . BBTLBO converges to the threshold except for functions , , and . From the results of total average FEs, BBTLBO converges faster than other algorithms on all unimodal functions and the majority of multimodal functions except for functions , , , and . The acceleration rates between BBTLBO and other algorithms are mostly 10 for functions , , , , , , and . From the results of total average SR, BBTLBO achieves the highest SR for those test functions of which BBTLBO successfully converges to the threshold value. It can be concluded that the BBTLBO has a good performance of convergence speed and successful rate (SR) of the solutions for test functions in this paper.
5.5. Comparison of BBTLBO with DE Variants, PSO Variants, and Some TLBO Variants
In this section, we compared the performance of BBTLBO with other optimization algorithms, including jDE , SaDE , PSOcfLocal , PSOwFIPS , and TLBO [8, 9]. In our experiment, the maximal FEs are used as the stopping criterion of all algorithms, namely, 40,000 for all test functions. The results are shown in Table 5 in terms of the mean optimum solution and the standard deviation of the solutions obtained in the 50 independent runs by each algorithm on 20 test functions, where “” summarizes the competition results among BBTLBO and other algorithms. The best results among the algorithms are shown in boldface.
The comparisons in Table 5 show that that all algorithms perform well for , , and . Although SaDE outperforms BBTLBO on , PSOcfLocal outperforms BBTLBO on and PSOwFIPS outperforms BBTLBO on and , and BBTLBO offers the highest accuracy on functions , , , , , , , and . “” shows that BBTLBO offers well accuracy for the majority of test functions in this paper. Table 5 concludes that BBTLBO has a good performance of the solution accuracy for all unimodal optimization problems and most complex multimodal optimization problems.
6. Two Real-World Optimization Problems
In this section, to show the effectiveness of the proposed method, the proposed BBTLBO algorithm is applied to estimate parameters of two real-world problems.
6.1. Nonlinear Function Approximation
The artificial neural network trained by our BBTLBO algorithm is a three-layer feed-forward network and the basic structure of the proposed scheme is depicted in Figure 5. The inputs are connected to all the hidden units, which in turn all connected to all the outputs. The variables consist of neural network weights and biases. Suppose a three-layer forward neural network architecture with input units, hidden units, and output units, and the number of the variables is shown as follows:
For neural network training, the aim is to find a set of weights with the smallest error measure. Here the objective function is the mean sum of squared errors (MSE) over all training patterns which is shown as follows: where is the number of training data set, is the number of output units, is desired output, and is output inferred from neural network.
In this example, a three-layer feed-forward ANN with one input unit, five hidden units, and one output unit is constructed to model the curve of a nonlinear function which is described by the following equation :
In this case, activation function used in the output layer is the sigma function and activation function used in the output layer is linear. The number (dimension) of the variables is 16 for BBTLBO-based ANN. In order to train the ANN, 200 pairs of data are chosen from the real model. For each algorithm, 50 runs are performed. The other parameters are the same as those of the previous investigations. The results are shown in Table 6 in terms of the mean MSE and the standard deviation obtained in the 50 independent runs for three methods. Figure 6 shows the predicted time series for training and test using different algorithms. It can conclude that the approximation achieved by BBTLBO has good performance.
6.2. Tuning of PID Controller
The continuous form of a discrete-type PID controller with a small sampling period is described as follows : where is the controlled output, respectively. is the error signal, and are the reference signal and the system output, and , , and represent the proportional, integral and derivate gains, respectively.
For an unknown plant, the goal of this problem is to minimize the integral absolute error (IAE), which is given as follow [32, 33]: where and are used to represent the system error and the control output at time , is the rising time, and ( = 1, 2, 3) are weight coefficients.
To avoid overshooting, a penalty value is adopted in the cost function. That is, once overshooting occurs, the value of overshooting is added to the cost function, and the cost function is given as follows [32, 33]: where is a coefficient and , , and is the output of the controlled objective.
In our simulation, the formulas for the plant examined are given as follows :
The system sampling time is second and the control value is limited in the range of . Other relevant system variables are , , and . The weight coefficients of the cost function are set as , , and in this example.
In the simulations, the step response of PID control system tuned by the proposed BBTLBO is compared with that tuned by the standard genetic algorithm (GA) and the standard PSO (PSO). The population sizes of GA, PSO, and BBTLBO are 50, and the corresponding maximum numbers of iterations are 50, 50, and 50, respectively. In addition, the crossover rate is set as 0.90 and the mutation rate is 0.10 for GA.
The optimal parameters and the corresponding performance values of the PID controllers are listed in Table 7 and the corresponding performance curves and step responses curves are given in Figures 7 and 8. It can be seen from Figure 7 and Table 7 that the PID controller tuned by BBTLBO has the minimum cost function and CPU time. Although PID controllers tuned by PSO have a smaller peak time and rise time, their maximum overshoots are much larger than the overshoot tuned by BBTLBO. It concludes that the PID controller tuned by the BBTLBO could perform the best control performance in the simulations.
In this paper, TLBO has been extended to BBTLBO which uses the hybridization of the learning strategy in the standard TLBO and Gaussian sampling learning to balance the exploration and the exploitation in teacher phase and uses a modified mutation operation so as to eliminate the duplicate learners in learner phase. The proposed BBTLBO algorithm is utilized to optimize 20 benchmark functions and two real-world optimization problems. From the analysis and experiments, the BBTLBO algorithm significantly improves the performance of the original TLBO, although it needs to spend more CPU time than the standard TLBO algorithm in each generation. From the results compared with other algorithms on the 20 chosen test problems, it can be observed that the BBTLBO algorithm has good performance by using neighborhood search more effectively to generate better quality solutions, although the BBTLBO algorithm does not always have the best performance in all experiments cases of this paper. It can be also observed that the BBTLBO algorithm gives the best performance on two real-world optimization problems compared with other algorithms in the paper.
Further work includes research into neighborhood search based on different topological structures. Moreover, the algorithm may be further applied to constrained, dynamic, and noisy single-objective and multiobjective optimization domain. It is expected that BBTLBO will be used to more real-world optimization problems.
Conflict of Interests
The authors declare that there is no conflict of interests regarding the publication of this paper.
This research was partially supported by the National Natural Science Foundation of China (61100173, 61100009, 61272283, and 61304082). This work is partially supported by the Natural Science Foundation of Anhui Province, China (Grant no. 1308085MF82), and the Doctoral Innovation Foundation of Xi’an University of Technology (207-002J1305).
- D. E. Goldberg, Genetic Algorithms in Search Optimization and Machine Learning, Addison-Wesley, Reading, Mass, USA, 1989.
- L. C. Jiao and L. Wang, “A novel genetic algorithm based on immunity,” IEEE Transactions on Systems, Man, and Cybernetics A: Systems and Humans, vol. 30, no. 5, pp. 552–561, 2000.
- R. Storn and K. Price, “Differential evolution: a simple and efficient Heuristic for global optimization over continuous spaces,” Journal of Global Optimization, vol. 11, no. 4, pp. 341–359, 1997.
- M. Dorigo and T. Stutzle, Ant Colony Optimization, MIT Press, 2004.
- J. Kennedy and R. Eberhart, “Particle swarm optimization,” in Proceedings of the IEEE International Conference on Neural Networks, pp. 1942–1948, December 1995.
- D. Karaboga and B. Basturk, “On the performance of artificial bee colony (ABC) algorithm,” Applied Soft Computing Journal, vol. 8, no. 1, pp. 687–697, 2008.
- D. Simon, “Biogeography-based optimization,” IEEE Transactions on Evolutionary Computation, vol. 12, no. 6, pp. 702–713, 2008.
- R. V. Rao, V. J. Savsani, and D. P. Vakharia, “Teaching-learning-based optimization: a novel method for constrained mechanical design optimization problems,” CAD Computer Aided Design, vol. 43, no. 3, pp. 303–315, 2011.
- R. V. Rao, V. J. Savsani, and D. P. Vakharia, “Teaching-learning-based optimization: an optimization method for continuous non-linear large scale problems,” Information Sciences, vol. 183, no. 1, pp. 1–15, 2012.
- R. V. Rao, V. J. Savsani, and D. P. Vakharia, “Teaching-learning-based optimization algorithm for unconstrained and constrained real-parameter optimization problems,” Engineering Optimization, vol. 44, no. 12, pp. 1447–1462, 2011.
- V. Toĝan, “Design of planar steel frames using teaching-learning based optimization,” Engineering Structures, vol. 34, pp. 225–232, 2012.
- R. V. Rao and V. Patel, “An elitist teaching-learning-based optimization algorithm for solving complex constrained optimization problems,” International Journal of Industrial Engineering Computations, vol. 3, pp. 535–560, 2012.
- S. O. Degertekin and M. S. Hayalioglu, “Sizing truss structures using teaching-learning-based optimization,” Computers and Structures, vol. 119, pp. 177–188, 2013.
- R. V. Rao and V. Patel, “An improved teaching-learning-based optimization algorithm for solving unconstrained optimization problems,” Scientia Iranica, vol. 20, no. 3, pp. 710–720, 2013.
- R. V. Rao and V. Patel, “Multi-objective optimization of combined Brayton and inverse Brayton cycles using advanced optimization algorithms,” Engineering Optimization, vol. 44, no. 8, pp. 965–983, 2011.
- T. Niknam, F. Golestaneh, and M. S. Sadeghi, “Theta-multi-objective teaching-learning-based optimization for dynamic economic emission dispatch,” IEEE Systems Journal, vol. 6, no. 2, pp. 341–352, 2012.
- R. V. Rao and V. Patel, “Multi-objective optimization of heat exchangers using a modified teaching-learning-based optimization algorithm,” Applied Mathematical Modelling, vol. 37, no. 3, pp. 1147–1162, 2013.
- M. Clerc and J. Kennedy, “The particle swarm-explosion, stability, and convergence in a multidimensional complex space,” IEEE Transactions on Evolutionary Computation, vol. 6, no. 1, pp. 58–73, 2002.
- F. van den Bergh and A. P. Engelbrecht, “A study of particle swarm optimization particle trajectories,” Information Sciences, vol. 176, no. 8, pp. 937–971, 2006.
- J. Kennedy, “Bare bones particle swarms,” in Proceedings of the Swarm Intelligence Symposium (SIS '03), pp. 80–87, 2003.
- M. G. H. Omran, A. P. Engelbrecht, and A. Salman, “Bare bones differential evolution,” European Journal of Operational Research, vol. 196, no. 1, pp. 128–139, 2009.
- H. Wang, S. Rahnamayan, H. Sun, and M. G. H. Omran, “Gaussian bare-bones differential evolution,” IEEE Transactions on Cybernetics, vol. 43, no. 2, pp. 634–647, 2013.
- X. H. Hu and R. Eberhart, “Multiobjective optimization using dynamic neighborhood particle swarm optimization,” in Proceedings of the Congress on Evolutionary Computation, pp. 677–1681, 2002.
- M. G. Omran, A. P. Engelbrecht, and A. Salman, “Using the ring neighborhood topology with self-adaptive differential evolution,” in Advances in Natural Computation, pp. 976–979, Springer, Berlin, Germany, 2006.
- X. Li, “Niching without niching parameters: particle swarm optimization using a ring topology,” IEEE Transactions on Evolutionary Computation, vol. 14, no. 1, pp. 150–169, 2010.
- I. Maruta, T. H. Kim, D. Song, and T. Sugie, “Synthesis of fixed-structure robust controllers using a constrained particle swarm optimizer with cyclic neighborhood topology,” Expert Systems with Applications, vol. 40, no. 9, pp. 3595–3605, 2013.
- J. Kennedy and R. Mendes, “Population structure and particle swarm performance,” in Proceedings of the International Conference on Evolutionary Computation, pp. 1671–1676, Honolulu, Hawaii, USA, 2002.
- J. Brest, S. Greiner, B. Bošković, M. Mernik, and V. Zumer, “Self-adapting control parameters in differential evolution: a comparative study on numerical benchmark problems,” IEEE Transactions on Evolutionary Computation, vol. 10, no. 6, pp. 646–657, 2006.
- A. K. Qin, V. L. Huang, and P. N. Suganthan, “Differential evolution algorithm with strategy adaptation for global numerical optimization,” IEEE Transactions on Evolutionary Computation, vol. 13, no. 2, pp. 398–417, 2009.
- R. Mendes, J. Kennedy, and J. Neves, “The fully informed particle swarm: simpler, maybe better,” IEEE Transactions on Evolutionary Computation, vol. 8, no. 3, pp. 204–210, 2004.
- F. Herrera and M. Lozano, “Gradual distributed real-coded genetic algorithms,” IEEE Transactions on Evolutionary Computation, vol. 4, no. 1, pp. 43–62, 2000.
- J. Liu, Advanced PID Control and MATLAB Simulation, Electronic Industry Press, 2003.
- J. Zhang, J. Zhuang, H. Du, and S. Wang, “Self-organizing genetic algorithm based tuning of PID controllers,” Information Sciences, vol. 179, no. 7, pp. 1007–1017, 2009.
- R. Haber-Haber, R. Haber, M. Schmittdiel, and R. M. del Toro, “A classic solution for the control of a high-performance drilling process,” International Journal of Machine Tools and Manufacture, vol. 47, no. 15, pp. 2290–2297, 2007.