Advances in Architectures, Big Data, and Machine Learning Techniques for Complex Internet of Things SystemsView this Special Issue
Research Article | Open Access
Qinghua Gu, Xuexian Li, Song Jiang, "Hybrid Genetic Grey Wolf Algorithm for Large-Scale Global Optimization", Complexity, vol. 2019, Article ID 2653512, 18 pages, 2019. https://doi.org/10.1155/2019/2653512
Hybrid Genetic Grey Wolf Algorithm for Large-Scale Global Optimization
Most real-world optimization problems tackle a large number of decision variables, known as Large-Scale Global Optimization (LSGO) problems. In general, the metaheuristic algorithms for solving such problems often suffer from the “curse of dimensionality.” In order to improve the disadvantage of Grey Wolf Optimizer when solving the LSGO problems, three genetic operators are embedded into the standard GWO and a Hybrid Genetic Grey Wolf Algorithm (HGGWA) is proposed. Firstly, the whole population using Opposition-Based Learning strategy is initialized. Secondly, the selection operation is performed by combining elite reservation strategy. Then, the whole population is divided into several subpopulations for cross-operation based on dimensionality reduction and population partition in order to increase the diversity of the population. Finally, the elite individuals in the population are mutated to prevent the algorithm from falling into local optimum. The performance of HGGWA is verified by ten benchmark functions, and the optimization results are compared with WOA, SSA, and ALO. On CEC’2008 LSGO problems, the performance of HGGWA is compared against several state-of-the-art algorithms, CCPSO2, DEwSAcc, MLCC, and EPUS-PSO. Simulation results show that the HGGWA has been greatly improved in convergence accuracy, which proves the effectiveness of HGGWA in solving LSGO problems.
Large-Scale Global Optimization (LSGO) is widely applied in practical engineering problems, such as large-scale job shop scheduling problem , large-scale vehicle routing problems , and reactive power optimization of large-scale power system . It is a very important and challenging task in the optimization domain and usually involves thousands of decision variables. As the number of dimensions increases, the complexity of the problem increases exponentially, so it is very difficult to solve. Metaheuristic algorithms are a class of intelligent optimization algorithms inspired by biological activities or physical principles. In recent years, various metaheuristic algorithms such as genetic algorithms, particle swarm optimization algorithms, artificial bee colony algorithms , and differential evolution algorithms  have been applied to a variety of large-scale global optimization problems. However, due to the existence of “curse of dimensionality”, it is difficult for general metaheuristic algorithms to find the optimal solution of LSGO problems. In order to solve LSGO problems better, many valuable attempts have been made in recent years by using metaheuristic algorithms.
In general, solving large-scale problems can be considered in two ways. First, from the perspective of the problem itself, the solution process can be made more efficient by simplifying the problem. Secondly, from the perspective of the method, the optimal solution to the problem can be solved with greater possibility by improving the performance of the solution algorithm. First of all, in terms of the problem itself, a large number of decision variables are the main cause of the complexity of the problem . In the famous book A Discourse on Method , Descartes pointed out that it is necessary to study complex problems and decompose them into a number of relatively simple small problems and then solve them one by one. He called it a “divide and conquer” strategy, which is to solve the whole problem by decomposing the original large-scale problem into a set of smaller and simpler subproblems which are more manageable and easier to solve and then solve each subproblem independently. This method based on “divide and conquer” strategy is also called Cooperative Coevolution (CC), and its effectiveness for solving large-scale optimization problems has been demonstrated in many classical optimization methods. But a major difficulty in applying CC is the choice of a good decomposition strategy, because different decomposition strategies may lead to different optimization effects. Due to the diversity of problem-solving, there may be very large differences between different problems. Therefore, it is necessary to study the structure inherent in the problem and analyze the relationship between the variables in order to find a suitable solution.
In addition, from an algorithmic perspective, two or more distinct methods when combined together in a synergistic manner can enhance the problem-solving capability of the derived hybrid . EAs hybridized with local search algorithms have been successful within function optimization domain. These kinds of EAs are often named as memetic algorithms (MAs) . The optimization performance of the algorithm for large-scale optimization problems is improved by different optimization strategies such as designing new mutation operators, dynamic neighborhood search strategies, multiple classifier , and Opposition-Based Learning . Similar to evolutionary computation, the swarm intelligence (SI) is a kind of optimization algorithm inspired by the biological foraging or hunting behavior in nature, which simulates the intelligent behavior of insects, bird groups, ant colonies, or fish schools. Inspired by the grey wolf population hierarchy and hunting behavior, Mirjalili proposed another swarm intelligence optimization algorithm, grey wolf optimization algorithm (GWO) in 2014 . The GWO algorithm has the advantages of less control parameters and fast convergence and it has been applied to various optimization fields, such as medical image fusion , multiple input multiple output power systems , and job shop scheduling problem . In addition, the proposal of different prediction techniques  also provides reference for the study of large-scale optimization problems.
To direct at the large-scale global optimization problems, this paper improves the HGGWA algorithm proposed in literature  by adding a parameter nonlinear adjustment strategy, which further improves the global convergence and solution accuracy. The crossover operation of the HGGWA algorithm divides the whole population into several subpopulations by referring to the idea of cooperative coevolution, and then independently evolves each subpopulation separately. More specifically, this paper has the following research objectives:
To review the current literature on solving large-scale global optimization problems and to analyze the existing problems in the research
To improve the Hybrid Genetic Grey Wolf Algorithm (HGGWA) by adding the nonlinear adjustment strategy of parameter, which further improves the global convergence and solution accuracy for solving large-scale global optimization problems
To analyze the global convergence and computational complexity of the improved HGGWA algorithm and to verify the effectiveness of HGGWA for solving large-scale global optimization problems by several different numerical experiments
The remainder of this paper is organized as follows: in Section 2, a review of Large-Scale Global Optimization (LSGO) and various solving strategies is given. Then, Section 3 briefly describes the Grey Wolf Optimizer (GWO). In Section 4, the principles and steps of the improved algorithm (HGGWA) are introduced. Section 5 illustrates and analyzes the experimental results. Finally, the conclusions are drawn in Section 6.
2. Literature Review
2.1. Decomposition-Based CC Strategy
Large-scale global optimization problems are mainly solved in the following two ways. One is Cooperative Coevolution (CC) with problem decomposition strategy. The idea of decomposition was first proposed by Potter and De Jong [18, 19] in 1994. They designed two Cooperative Coevolutionary Genetic Algorithms (CCGA-1, CCGA-2) to improve the global convergence of GA. The CC strategy is to reduce the scale of the high-dimensional problem by dividing the whole solution space into multiple subspaces and then use multiple subpopulations to achieve the cooperative coevolution. Table 1 summarizes proposed major variants of the CC method. It can be divided into the static grouping-based CC methods and the dynamic grouping-based CC methods.
The static grouping-based CC methods divide the population into multiple fixed-scale subpopulations. The Cooperative Coevolutionary Genetic Algorithm (CCGA) proposed by Potter and De Jong is the first attempt to combine the idea of Cooperative Coevolution with metaheuristic algorithm. They decomposed an n-dimensional problem into n 1-dimensional subproblems each of which is optimized by a separated GA. Van den Bergh et al.  firstly proposed two improved PSO by integrating cooperative coevolution strategy into PSO, called CPSO-SK and CPSO-HK, which divides an n-dimensional problem into k s-dimensional subproblems. Similarly, Mohammed EI-Abd  presented two cooperative coevolutionary ABC algorithms, namely, CABC-S and CABC-H. Shi et al.  applied the cooperative coevolution strategy to the differential evolution algorithm, called the cooperative coevolutionary differential evolution (CCDE). The algorithm partitions high-dimensional solution space into two equal-sized subcomponents but it does not perform well as the dimensionality increases. For a fully separable problem with no interrelationship between variables, each variable can be optimized independently to obtain the optimal solution for the entire problem. However, for the nonseparable problem of the relationship between variables, the optimization effect of the fixed-scale grouping strategy will be worse, and even the optimal solution will not be obtained. Therefore, many scholars began to study the decomposition strategy that can solve the nonseparable problems.
The dynamic grouping-based CC methods dynamically adjust the size of the subpopulation to accommodate different types of large-scale optimization problems. They are efficient to handle nonseparable problems. Yang et al.  proposed a new cooperative coevolution framework that is capable of optimizing large-scale nonseparable problems. It uses a random grouping strategy to divide the problem solution space into multiple subspaces of different sizes. Li et al.  embedded a random grouping strategy and an adaptive weighting mechanism into the particle swarm optimization algorithm and proposed a cooperative coevolutionary particle swarm optimization (CCPSO) algorithm. The results of the random grouping method show the effective performance to solve the scalable nonseparable benchmark function (up to 1000D). However, as the number of interaction variables increases, its performance becomes invalid . Therefore, some scholars consider adding a priori knowledge of the problem in the process of algorithm evolution and propose many learning-based dynamic grouping strategies. Ray et al.  discussed the effects of problem size, the number of subpopulations, and the number of iterations on some separable and nonseparable problems in cooperative coevolutionary algorithms. Then, a Cooperative Coevolutionary Algorithm with Correlation based Adaptive Variable Partitioning (CCEA-AVP) is proposed to deal with scalable nonseparable problems. Chen et al.  proposed a CC method with Variable Interaction Learning (CCVIL) which is able to adaptively change group sizes. Considering that there is no prior knowledge that cannot decompose the problem, Omidvar et al.  proposed an automatic decomposition strategy, which can keep the correlation between decision variables at the minimum.
2.2. Nondecomposition Strategy
The other is nondecomposition methods. Different from the decomposition-based strategy, the nondecomposition methods mainly study the algorithm itself and improve the corresponding operator in order to improve the performance of the algorithm for solving large-scale global optimization problems. Nondecomposition-based methods mainly include swarm intelligence, evolutionary computation, and local search-based approaches. Hsieh et al.  added variable particles and information sharing mechanism to the basic particle swarm optimization algorithm and proposed a variation called the Efficient Population Utilization Strategy for Particle Swarm Optimizer (EPUS-PSO). A modified PSO algorithm with a new velocity updating method, as the rotated particle swarm, was proposed in literature . It transforms coordinates and uses information from other dimensions to maintain the diversity of each dimension. Fan et al.  proposed a novel particle swarm optimization approach with dynamic neighborhood that is based on kernel fuzzy clustering and variable trust region methods (called FT-DNPSO) for large-scale optimization. In terms of evolutionary computation, Hedar et al.  modified genetic algorithm with new strategies of population partitioning and space reduction for high-dimensional problems. In order to improve the performance of differential evolution algorithm, Zhang et al.  proposed an adaptive differential evolution algorithm (called JADE) using a new mutation and adaptive control parameter strategy. Local search-based approaches have been recognized as an effective algorithm framework for solving optimization problems. Hvattum et al.  introduced a new direct search method, based on Scatter Search, designed to remedy the lack of a good derivative-free method for solving problems of high dimensions. Liu et al.  improved the local search depth parameter in the memetic algorithm and proposed an adaptive local search depth (ALSD) strategy, which can arrange the local search computing resources according to its performance dynamically.
2.3. Related Work of GWO
Grey Wolf Optimizer (GWO) is a recently proposed swarm intelligence algorithm inspired by the social hierarchy and hunting behavior of wolves. It has the advantages of less control parameters, high solution accuracy, and fast convergence speed. Compared with other classical metaheuristic algorithms such as genetic algorithm (GA), particle swarm optimization (PSO), and gravitational search algorithm (GSA), the GWO shows powerful exploration capability and better convergence characteristics. Owing to its simplicity and ease of implementation, GWO has gained significant attention and has been applied in solving many practical optimization problems since its invention.
Recently, many scholars have applied the GWO to different optimization problems and also proposed many varieties in order to improve the performance of GWO. Saremi et al.  introduced a dynamic population updating mechanism in the GWO, removing individuals with poor fitness in the population and regenerating new individuals to replace them in order to enhance the exploration ability of GWO. However, the diversity of grey population was reduced by this strategy, which leads to the algorithm being local optimum. Zhu et al.  integrated the idea of differential evolution into the GWO, which prevented the early stagnation of the GWO algorithm and accelerated the convergence speed. However, it fails to converge in high-dimensional problems. Kumar et al.  designed a novel variant of GWO for numerical optimization and engineering design problems. They utilized the prey weight and astrophysics-based learning concept to enhance the exploration ability. However, this method also does not show good performance in solving high-dimensional function problems. In the application of the GWO algorithm, Guha et al.  applied the grey wolf algorithm to the load control problem in large-scale power systems. In literature , a multiobjective discrete grey wolf algorithm is proposed for the uniqueness of welding job scheduling problem. The results show that the algorithm can solve the scheduling problem in practical engineering. Emary et al.  proposed a new binary grey wolf optimization algorithm for the selection of the best feature subset.
The swarm intelligence optimization algorithm has natural parallelism, so it has advantages in solving large-scale optimization problems. In order to study the ability of grey wolf optimization algorithm to solve high-dimensional functions, Long et al.  proposed a nonlinear adjustment strategy for convergence factors, which balanced well the exploration and exploitation of the algorithm. But this strategy also does not solve the high-dimensional function optimization problem well. In 2017, Gupta et al.  studied the performance of the GWO algorithm for solving 100 to 1000-dimensional function optimization problems. The results indicate that GWO is a powerful nature-inspired optimization algorithm for large-scale problems, except Rosenbrock function.
In summary, the decomposition-based CC strategy has a good effect in solving the nonseparable problem, but it cannot solve the imbalance problem well, and it still needs to be improved in the accuracy of the solution. In addition, in the method based on nondecomposition strategy, although the performance of many algorithms has been greatly improved, but the research on hybrid algorithms is still rare. Therefore, this paper combines three genetic operators of selection, crossover, and mutation with GWO algorithm and proposes a Hybrid Genetic Grey Wolf Algorithm, (HGGWA). The main improvement strategies of the algorithm are as follows: initialize the population based on the Opposition-Based Learning strategy; the nonlinear adjustment of parameters effectively balances exploration and exploitation capabilities while accelerating the convergence speed of the algorithm; select the population through a combination of elite retention strategy and gambling roulette; based on the method of dimensionality reduction and population division, the whole population is divided into multiple subpopulations for cross-operation to increase the diversity of the population; perform mutation operations on elite individuals in the population to prevent the algorithm from falling into local optimum.
3. Grey Wolf Optimization Algorithm
3.1. Mathematical Models
In the GWO algorithm , the position of α wolf, β wolf, and δ wolf is the fittest solution, the second-best solution, and the third-best solution, respectively. And the positions of the ω wolves are the remaining candidate solutions. In the whole process of searching for prey in the space, the ω wolves gradually update their positions according to the optimal position of the α, β, and δ wolves. So they can gradually approach and capture the prey, that is, to complete the process of searching for the optimal solution. The position updating imaginary diagram in GWO is shown in Figure 1.
The grey wolves position update formula in this process is as follows:
where indicates the distance between the grey wolf individual and the prey, indicates the current iteration, and are the coefficient vectors, is a convergence coefficient used to balance the exploration and the exploitation, is used to simulate the effects of nature, is the position vector of the prey, and indicates the position vector of a grey wolf.
The coefficient vectors and are calculated as follows:
where components of are linearly decreased from 2 to 0 over the course of iterations and and are random vectors in . At the same time, the fluctuation range of the coefficient vector gradually decreases as decreases.
where , , indicate the distance between the α, β, δ wolves and the prey, respectively. , , represent the positional components of the remaining grey wolves based on the positions of α, β, and δ wolves, respectively. represents the position vector after the grey wolf is updated.
3.2. Basic Process of GWO
Step 1. Randomly generate an initial population, and initialize parameters , , and .
Step 2. Calculate the fitness of the grey wolf individual, and save the positions of the first three individuals with the highest fitness as , , and .
Step 4. Calculate the fitness of individuals in the new population, and update , , and .
4. Hybrid Genetic Grey Wolf Algorithm
This section describes the details of HGGWA, the hybrid algorithm proposed in this paper. Firstly, the initial population strategy based on Opposition-Based Learning is described in detail in Section 4.1. Secondly, Section 4.2 introduces the nonlinear adjustment strategy for parameter . Then three operations of selection, crossover, and mutation are carried out in Sections 4.3–4.5, respectively. Finally, the time complexity of HGGWA is analyzed in Section 4.6.
4.1. Initial Population Strategy Based on Opposition-Based Learning
For the swarm intelligence optimization algorithm based on population iteration, the diversity of initial populations lays the foundation for the efficiency of the algorithm. The better diversity of the population will reduce the computation time and improve the global convergence of the algorithm . However, like other algorithms, the GWO algorithm also uses random initialization when generating populations, which will have a certain impact on the search efficiency of the algorithm. Opposition-Based Learning is a strategy proposed by Tizhoosh  to improve the efficiency of algorithm search. It uses the opposite points of known individual positions to generate new individual positions, thereby increasing the diversity of search populations. Currently, the Opposition-Based Learning strategy has been successfully applied to multiple swarm optimization algorithms [66, 67]. The opposite point is defined as follows.
Assume that there is an individual in the population P, then the opposite point of the element on each dimension is , where and are the lower bound and the upper of the th dimension, respectively. According to the above definition, the Opposition-Based Learning strategy initializes the population as follows:
(a) Randomly initialize the population P, calculate the opposite point of each individual position , and all the opposite points constitute the opposition population .
(b) Calculate the fitness of each individual in the initial population P and the opposition population , and arrange them in descending order.
(c) Select the top N grey wolf individuals with the highest fitness as the final initial population.
4.2. Parameters Nonlinear Adjustment Strategy
It is significantly important how to balance the exploration and exploitation for swarm intelligence algorithms. In the early iteration of the algorithm, the powerful exploration capability is beneficial for the algorithm to expand the search range, so that the algorithm can search for the optimal solution with greater probability. In the later stage of the algorithm’s iteration, the effective exploitation ability can speed up the optimization process of the algorithm and improve the convergence accuracy of the final solution. Therefore, only when the swarm optimization algorithm can better coordinate its global exploration and local exploitation capabilities can it have strong robustness and faster convergence speed.
It can be seen from (2) and (3) that the dynamic change of the parameter plays a crucial role in the algorithm. In the basic GWO, the linear decrement strategy of parameter does not reflect the actual convergence process of the algorithm. Therefore, the algorithm does not balance the exploration and exploitation capabilities well, especially when solving large-scale multimodal function problems. In order to dynamically adjust the global exploration and local exploitation process of the algorithm, this paper proposes a nonlinear adjustment strategy for the parameter a, which is beneficial for HGGWA to search for the optimal solution. The improved parameter nonlinear adjustment equation is as follows:
where and are the initial value and the end value of the parameter , is the current iteration index, is the maximum number of iterations, and is nonlinear adjustment coefficient.
Thus, the variation range of the convergence factor A is controlled by the nonlinear adjustment of the parameter . When , the grey wolf population expands the search range to find better prey, which corresponds to the global exploration of the algorithm; when , the whole population shrinks the search range; thus, an encirclement is formed around the prey to complete the final attacking against the prey, which corresponds to the local exploitation process of the algorithm. The whole process has a positive effect on balancing global search and local search, which is beneficial to improve the accuracy of the solution and further accelerate the convergence speed of the algorithm.
4.3. Selection Operation Based on Optimal Retention
For the intelligent optimization algorithm based on population evolution, the evolution of each generation of population directly affects the optimization effect of the algorithm. In order to inherit the superior individuals in the paternal population to the next generation without being destroyed, it is necessary to preserve the good individuals directly. The optimal retention selection strategy is an effective way to preserve good individuals in genetic algorithms. This paper integrates them into the GWO to improve the efficiency of the algorithm. Assuming that the current population is P(), the fitness of the individual is . The specific operation is as follows:
(a) Calculate the fitness of each individual and arrange them in descending order.
(b) Selecting the individuals with the highest fitness to copy directly into the next generation of population.
(c) Calculate the total fitness of the remaining individuals and the probability that each individual is selected.
(d) Calculate the cumulative fitness value for each individual, and then the selection operation is performed in the manner of the bet roulette until the number of individuals in the children population is consistent with the parent population.
4.4. Crossover Operation Based on Population Partitioning Mechanism
Due to the decrease of population diversity in the late evolution, the GWO algorithm is easy to fall into local optimum for solving the large-scale high-dimensional optimization problem, in order to overcome the problems caused by large-scale and complexity and to ensure that the algorithm searches for all solution spaces.
In the HGGWA algorithm, the whole population P is divided into subpopulations . The optimal size of each subpopulation was tested to be 5 × 5. Individuals of each subpopulation were cross-operated to increase the diversity of the population. The specific division method is shown below.
In genetic algorithms, cross-operation has a very important role and is the main way to generate new individuals. In the improved algorithm of this paper, each individual in the subpopulation is cross-operated in a linear crossover manner. Generating a corresponding random number (0,1) for each individual in the subpopulation. When the random number is less than the crossover probability , the corresponding individual is paired for cross-operation.
An example is as follows: Crossover (p1, p2)
(a) Generate a random number .
(b) Two children , are generated by two parents and :
4.5. Mutation Operation for Elite Individuals
Due to the existence of the selection operation based on the optimal preservation strategy, the grey wolf individuals in the whole population are concentrated in a small optimal region in the later stage of the iterative process, which easily leads to the loss of population diversity. If the current optimal individual is a locally optimal solution, then the algorithm is easy to fall into local optimum, especially when solving high-dimensional multimodal functions. To this end, this paper introduces the mutation operator in the HGGWA algorithm to perform mutation operations on elite individuals in the population. The specific operations are as follows.
Assume that the optimal individual is and the mutation operation is performed on with the mutation probability . That is, select a gene from the optimal individual with probability , instead of the gene with a random number between upper and lower bounds to generate a new individual . The specific operation is as follows:
where is a random number in and and are the lower and upper bounds of the individual , respectively.
4.6. Pseudo Code of the HGGWA and Time Complexity Analysis
This section describes how to calculate an upper bound for the total number of fitness evaluations (FE) required by HGGWA. As shown in Algorithm 1, the computational complexity of HGGWA in one generation is mainly dominated by the position updating of all search agents in line (10). The position updating requires a time complexity of O(Ndim)(N is the population size, dim is the dimensions of each benchmark function, and is the maximum number of iterations of HGGWA) to obtain the needed parameters and to finish the whole position updating progress of all search agents. In addition, the process of crossover operations is another time-consuming step in line (20). It needs a time complexity of O(2)( is the total number of subpopulations, =N/5, =dim/5) to complete the crossing of individuals in each subpopulations according to the crossover probability . It should be noted that this is only the worst case computational complexity. If there are only two individuals in each subpopulation that need to be crossed, then the minimum amount of computation is O(). Therefore, the maximum computational complexity caused by the two dominant processes in the algorithm is max in one generation, i.e., .
5. Numerical Experiments and Analysis
In this section, the proposed HGGWA algorithm will be evaluated on both classical benchmark functions  and the suite of test functions provided by CEC2008 Special Session on large-scale global optimization. The algorithms used for comparison include not only conventional EAs but also other CC optimization algorithms. Experimental results are provided to analyze the performance of HGGWA in the context of large-scale optimization problems. In addition, the sensitivity analysis of nonlinear adjustment coefficient and the global convergence of HGGWA are discussed in Sections 5.4 and 5.5, respectively.
5.1. Benchmark Functions and Performance Measures
In order to verify the ability of HGGWA algorithm to solve high-dimensional complex functions, 10 general high-dimensional benchmark functions (100D, 500D, 1000D) were selected for optimization test. Among them, are unimodal functions, which are used to test the local searchability of the algorithm; are multimodal functions, which are used to test the global searchability of the algorithm. These test functions have multiple local optima points with uneven distribution, nonconvexity, and strong oscillation, which are very difficult to converge to the global optimal solution, especially in the case of being high-dimensional. The specific characteristics of the 10 benchmark functions are shown in Table 2.
In order to show the optimization effect of the algorithm more clearly, this paper selects two indicators to evaluate the performance of the algorithm . The first is the solution accuracy (AC), which reflects the difference between the optimal results of the algorithm and the theoretical optimal value. In an experiment, if the final convergence result of the algorithm is less than the AC, then the optimization is considered successful. Assuming that the optimal value of a certain optimization is and the theoretical optimal value is , then the AC is calculated as follows:
The second is the successful ratio (SR), which reflects the proportion of the number of successful optimizations to the total number of optimizations under the condition that the algorithm has a certain accuracy. Assume that in the iterations of the algorithm, the number of successful optimizations is , then the SR is calculated as follows:
5.2. Results on Classical Benchmark Functions
In this section, a set of simulation tests were performed in order to verify the effectiveness of HGGWA. Firstly, the 10 high-dimensional benchmark functions in Table 2 are optimized by HGGWA algorithm proposed in this paper, and the results are compared with WOA, SSA, and ALO. In addition, the running time of several algorithms is compared under the given convergence accuracy.
5.2.1. Parameters Settings for the Compared Algorithms
In all experiments, the values of the common parameters, such as the maximum number of iterations (), the dimension of the functions (D), and the population sizes (N) were chosen the same. For all test problems, we focus on investigating the optimization performance of the proposed method on problems with D = 100, 500, and 1000. The maximum number of iterations is set to 1000, and the population size is set to 50. The parameter setting of WOA, SSA, and ALO is derived from the original literature [70–72]. The optimal parameter settings of the HGGWA algorithm proposed in this paper are shown in Table 3.
For each experiment of an algorithm on a benchmark function, 30 independent runs are performed to obtain a fair comparison among the different algorithms. The optimal value, the worst value, the average value, and the standard deviation of each algorithm optimization are recorded, and then the optimization successful ratio (SR) is calculated. All programs were coded in Matlab 2017b (Win64) and executed on a Lenovo computer with Intel (R) Core I5-6300HQ, 8G ROM, 2.30GHz under Windows 10 operating system.
5.2.2. Comparison with State-of-the-Art Metaheuristic Algorithms
In order to verify the efficiency of the HGGWA algorithm, several state-of-the-art metaheuristic algorithms recently proposed are compared with their optimization results. These algorithms are Whale Optimization Algorithm (WOA), Salp Swarm Algorithm (SSA), and Ant Lion Optimization Algorithm (ALO). The test was performed using the same 10 high-dimensional functions in Table 2, and the results of the comparison are shown in Table 4.
From Table 4, it can be known that the convergence accuracy and optimization successful ratio of the HGGWA algorithm are significantly higher than the other three algorithms for most of the test functions. In the case of 100 dimensions, HGGWA algorithm can converge to the global optimal solution for the other nine test functions at one time in addition to function 3. And the algorithm has a successful ratio of 100% for the nine functions under the condition of certain convergence accuracy. The convergence result of the WOA algorithm is better than the other three algorithms for functions and , and its robustness is excellent. With the increase of function dimension, the convergence precision of several algorithms decreases slightly, but the optimization results of HGGWA algorithm are better than those of WOA, SSA, and ALO. In general, the HGGWA algorithm exhibits better optimization effects than the WOA, SSA, and ALO algorithms in the 100-, 500-, and 1000-dimensional test functions, which proves the effectiveness of the HGGWA algorithm for solving high-dimensional complex functions.
Generally speaking, the HGGWA algorithm performs better than other algorithms in most test functions. In order to compare the convergence performance of the four algorithms more clearly, the convergence curves of the four algorithms on Sphere, Schwefel’s 2.22, Rastrigin, Griewank, Quartic, and Ackley functions with D=100 are shown in Figure 2. Sphere and Schwefel’s 2.22 represent the two most basic unimodal functions. Rastrigin and Griewank represent the two most basic multimodal functions. It is seen that HGGWA has higher precision and faster convergence speed than other algorithms.
(a) Sphere function
(b) Schwefel 2.22 function
(c) Rastrigin function
(d) Griewank function
(e) Quartic function
(f) Ackley function
5.2.3. Comparative Analysis of Running Time
To evaluate the actual runtime of the several compared algorithms, including SSA, WOA, GWO, and HGGWA, their average running times (in seconds: s) from 30 runs on function , , , , , and are plotted in Figure 3. The convergence accuracy of each test function has been described in Table 2.
(a) Comparative analysis of running time
(b) Comparative analysis of iterations
Figure 3 shows the convergence behavior of different algorithms. Each point on the plot was calculated by taking the average of 30 independent runs. As can be seen from Figure 3, several algorithms show different effects under the same calculation accuracy. For the functions , , , , , the GWO algorithm has the fastest convergence rate as a result of the advantages of the GWO algorithm itself. However, for the test functions and , the running time of the HGGWA algorithm is the shortest. Through further analysis, it is known that HGGWA algorithm adds a little computing time compared to GWO algorithm because of several improved strategies such as selection, crossover, and mutation. However, the convergence speed of HGGWA algorithm is still better than SSA and WOA algorithms. Overall, the convergence speed of the HGGWA algorithm performs better in several algorithms compared. Moreover, the running time of different test functions is very small, and it is kept between 1s and 3s, which shows that the HGGWA algorithm has excellent stability.
5.3. Results on CEC’2008 Benchmark Functions
This section provides an analysis of the effectiveness of HGGWA in terms of the CEC’2008 large-scale benchmark functions and a comparison with four powerful algorithms which have been proven to be effective in solving large-scale optimization problems. Experimental results are provided to analyze the performance of HGGWA for large-scale optimization problems.
We tested our proposed algorithms with CEC’2008 functions for 100, 500, and 1000 dimensions and the means of the best fitness value over 50 runs were recorded. While F1 (Shifted Sphere), F4 (Shifted Rastrigin), and F6 (Shifted Ackley) are separable functions, F2 (Schwefel Problem), F3 (Shifted Rosenbrock), F5 (Shifted Griewank), and F7 (Fast Fractal) are nonseparable, presenting a greater challenge to any algorithm that is prone to variable interactions. The performance of HGGWA is compared with other algorithms CCPSO2 , DEwSAcc , MLCC , and EPUS-PSO  for 100, 500, and 1000 dimensions, respectively. The maximum number of fitness evaluations (FEs) was calculated by the following formula, FEs = 5000×D, where D is the number of dimensions. In order to reflect the fairness of the comparison results, the optimization results of the other four algorithms are directly derived from the original literature. The specific comparison results are shown in Table 5.
Table 5 shows the experimental results and the entries shown in bold are significantly better than other algorithms. A general trend that can be seen is that HGGWA algorithm has a promising optimization effect for most of the 7 functions.
Generally speaking, HGGWA outperforms CCPSO2 on 6 out of 7 functions with 100 and 500 dimensions, respectively. Among them, the result of HGGWA is slightly worse than CCPSO2 for function F7 in the case of 100-dimensional; the two algorithms obtain similar optimization results for the function F6 in the case of 500-dimensional. In addition, at first glance it might seem that HGGWA and MLCC have similar performance. However, the HGGWA achieves better convergence accuracy than MLCC under the same number of iterations. In particular, the HGGWA algorithm’s optimization results are superior to DEwSAcc and EPUS-PSO in all dimensions. The result of HGGWA scaled very well from 100-D to 1000-D on F1, F4, F5, and F6, outperforming DEwSAcc and EPUS-PSO on all the cases.
5.4. Sensitivity Analysis of Nonlinear Adjustment Coefficient
This section mainly discusses the effect of nonlinear adjustment coefficients on the convergence performance of the algorithm. In the HGGWA algorithm, the role of parameter is to balance global exploration capabilities and local exploitation capabilities. In (12), the nonlinear adjustment coefficient is a key parameter and is mainly used to control the range of variation of the convergence factor. Therefore, we selected four different values for numerical experiments to analyze the effect of nonlinear adjustment coefficients on the performance of the algorithm. The four different values are 0.5,1,1.5,2, and the results of the comparison are shown in Table 6. Among them, black bold represents the best result of solving in several algorithms.
In general, the value of the nonlinear adjustment coefficient has no significant effect on the performance of the HGGWA algorithm. On closer inspection, one can see that besides the function , the optimization performance of the algorithm is optimal when the nonlinear adjustment coefficient is 0.5. But for function , = 1.5 is the best choice. For functions and , the HGGWA algorithm obtains the optimal solution 0 in the case of several different values of the coefficient . Therefore, for most test functions, the optimal value of the nonlinear adjustment factor is 0.5.
5.5. Global Convergence Analysis of HGGWA
In the HGGWA algorithm, the following four improved strategies are used to ensure the global convergence of the algorithm: initial population strategy based on Opposition-Based Learning; selection operation based on optimal retention; crossover operation based on population partitioning mechanism; and mutation operation for elite individuals.
The idea of solving the high-dimensional function optimization problem by HGGWA algorithm is as follows. Firstly, for the initial population strategy, the basic GWO generates the initial population in a random manner. However, it will have a great impact on the search efficiency of the algorithm for high-dimensional function optimization problems. The algorithm cannot effectively search the entire problem solution space if the initial grey wolf population is clustered in a small range. Therefore, the Opposition-Based Learning strategy is used to initialize the population, which can make the initial population evenly distributed in the solution space, thus laying a good foundation for the global search of the algorithm. Secondly, the optimal retention strategy is used to preserve the optimal individual of the parent population in the process of population evolution. As a result, the next generation of population can evolve in the optimal direction. In addition, the individuals with higher fitness are selected by the gambling roulette operation, which maintains the dominance relationship between individuals in the population. This dominance relationship makes the algorithm with good global convergence. Thirdly, the crossover operation is completed under the condition of population partitioning in order to achieve the purpose of dimension reduction and maintain the diversity of the population. Finally, the search direction of the grey wolf population is guided by the mutation operation of the elite individuals, which effectively prevents the algorithm from falling into the local optimal solution. All in all, through the improvement of these four different strategies, the HGGWA algorithm shows good global convergence in solving high-dimensional function optimization problems.
In order to overcome the shortcomings of GWO algorithm for solving large-scale global optimization problems which are easy to fall into local optimum, this paper proposes a Hybrid Genetic Grey Wolf Algorithm (HGGWA) by integrating three genetic operators into the algorithm. The improved algorithm initializes the population based on the Opposition-Based Learning strategy for improving the search efficiency of the algorithm. The strategy of population division based on cooperative coevolution reduces the scale of the problem, and the mutation operation of the elite individuals effectively prevents the algorithm from falling into the local optima. The performance of HGGWA has been evaluated using 10 classical benchmark functions and 7 CEC’2008 high-dimensional functions. From our experimental results, several conclusions can be drawn.
The results have shown that it is capable of grouping interacting variables with great accuracy for the majority of the benchmark functions. A comparative study among the HGGWA and other state-of-the-art algorithms was conducted and the experimental results showed that the HGGWA algorithm exhibits better global convergence whether it is solving separable functions or nonseparable functions.
By testing 10 classical benchmark functions, compared with the results of WOA, SSA, and ALO, the HGGWA algorithm has been greatly improved in convergence accuracy. In addition to the Schwefel problem and Rosenbrock function, the HGGWA algorithm achieves an 80%-100% successful ratio for all benchmark functions with 100 dimensions, which proves the effectiveness of HGGWA for solving high-dimensional functions.
By comparing with the optimization results on seven CEC’2008 large-scale global optimization problems, the results show that HGGWA algorithm achieves the global optimal value for the separable functions F1, F4, and F6, but the effect on other nonseparable problems is not very satisfactory. However, HGGWA algorithm still exhibits better global convergence than the other algorithms: CCPSO2, DewSAcc, MLCC, and EPUS-PSO on most of the functions.
By analyzing the running time of several algorithms, the results show that the convergence time of HGGWA algorithm has obvious advantages compared with SSA, WOA, etc. algorithms which are proposed recently. For the functions , , , , , and , the running time of the HGGWA algorithm is kept between 1s and 3s under the specified convergence accuracy, which shows the excellent stability of the HGGWA.
In the future, we are planning to investigate more efficient population partition mechanism to adapt to different nonseparable problems in the process of crossover operation. We are also interested in applying HGGWA to real-world problems to ascertain its true potential as a valuable optimization technique for large-scale optimization such as the setup of large-scale multilayer sensor networks  in the Internet of Things.
The data used to support the findings of this study are available from the corresponding author upon request.
Conflicts of Interest
The authors declare that there are no conflicts of interest regarding the publication of this paper.
Qinghua Gu and Xuexian Li contributed equally to this work.
This work was supported by National Natural Science Foundation of China (Grant Nos. 51774228 and 51404182), Natural Science Foundation of Shaanxi Province (2017JM5043), and Foundation of Shaanxi Educational Committee (17JK0425).
- R. Zhang and C. Wu, “A hybrid approach to large-scale job shop scheduling,” Applied Intelligence, vol. 32, no. 1, pp. 47–59, 2010.
- M. Gendreau and C. D. Tarantilis, Solving large-scale vehicle routing problems with time windows: The state-of-the-art[M]. Montreal: Cirrelt, 2010.
- M. B. Liu, S. K. Tso, and Y. Cheng, “An extended nonlinear primal-dual interior-point algorithm for reactive-power optimization of large-scale power systems with discrete control variables,” IEEE Power Engineering Review, vol. 22, no. 9, p. 56, 2002.
- I. Fister, I. Fister Jr., J. Brest, and V. Žumer, “Memetic artificial bee colony algorithm for large-scale global optimization,” in Proceedings of the IEEE Congress on Evolutionary Computation (CEC '12), pp. 1–8, IEEE, Brisbane, Australia, June 2012.
- A. Zamuda, J. Brest, B. Bošović, and V. Žumer, “Large scale global optimization using differential evolution with self-adaptation and cooperative co-evolution,” in Proceedings of the IEEE Congress on Evolutionary Computation (CEC '08), IEEE, June 2008.
- T. Weise, R. Chiong, and K. Tang, “Evolutionary optimization: pitfalls and booby traps,” Journal of Computer Science and Technology, vol. 27, no. 5, pp. 907–936, 2012.
- L. Lafleur, Descartes: Discourse On Method, Pearson Schweiz Ag, Zug, Switzerland, 1956.
- Y.-S. Ong, M. H. Lim, and X. Chen, “Memetic Computation—Past, Present & Future [Research Frontier],” IEEE Computational Intelligence Magazine, vol. 5, no. 2, pp. 24–31, 2010.
- N. Krasnogor and J. Smith, “A tutorial for competent memetic algorithms: model, taxonomy, and design issues,” IEEE Transactions on Evolutionary Computation, vol. 9, no. 5, pp. 474–488, 2005.
- S. Jiang, M. Lian, C. Lu, Q. Gu, S. Ruan, and X. Xie, “Ensemble prediction algorithm of anomaly monitoring based on big data analysis platform of open-pit mine slope,” Complexity, vol. 2018, Article ID 1048756, 13 pages, 2018.
- R. S. Rahnamayan, H. R. Tizhoosh, and M. M. A. Salama, “Opposition-based differential evolution,” IEEE Transactions on Evolutionary Computation, vol. 12, no. 1, pp. 64–79, 2008.
- S. Mirjalili, S. M. Mirjalili, and A. Lewis, “Grey wolf optimizer,” Advances in Engineering Software, vol. 69, pp. 46–61, 2014.
- E. Daniel, J. Anitha, K. K. Kamaleshwaran, and I. Rani, “Optimum spectrum mask based medical image fusion using gray wolf optimization,” Biomedical Signal Processing and Control, vol. 34, pp. 36–43, 2017.
- A. A. M. El-Gaafary, Y. S. Mohamed, A. M. Hemeida, and A.-A. A. Mohamed, “Grey wolf optimization for multi input multi output system,” Universal Journal of Communications and Network, vol. 3, no. 1, pp. 1–6, 2015.
- G. M. Komaki and V. Kayvanfar, “Grey wolf optimizer algorithm for the two-stage assembly flow shop scheduling problem with release time,” Journal of Computational Science, vol. 8, pp. 109–120, 2015.
- S. Jiang, M. Lian, C. Lu et al., “SVM-DS fusion based soft fault detection and diagnosis in solar water heaters,” Energy Exploration & Exploitation, 2018.
- Q. Gu, X. Li, C. Lu et al., “Hybrid genetic grey wolf algorithm for high dimensional complex function optimization,” Control and Decision, pp. 1–8, 2019.
- M. A. Potter and K. A. D. Jong, “A cooperative coevolutionary approach to function optimization,” in Parallel Problem Solving from Nature—PPSN III, vol. 866 of Lecture Notes in Computer Science, pp. 249–257, Springer, Berlin, Germany, 1994.
- M. A. Potter, The Design and Analysis of a Computational Model of Cooperative Coevolution, George Mason University, Fairfax, Va, USA, 1997.
- F. van den Bergh and A. P. Engelbrecht, “A cooperative approach to participle swam optimization,” IEEE Transactions on Evolutionary Computation, vol. 8, no. 3, pp. 225–239, 2004.
- M. El-Abd, “A cooperative approach to the artificial bee colony algorithm,” in Proceedings of the 2010 6th IEEE World Congress on Computational Intelligence, WCCI 2010 - 2010 IEEE Congress on Evolutionary Computation, CEC 2010, IEEE, Spain, July 2010.
- Y. Shi, H. Teng, and Z. Li, “Cooperative co-evolutionary differential evolution for function optimization,” in Proceedings of the 1st International Conference on Natural Computation (ICNC '05), Lecture Notes in Computer Science, pp. 1080–1088, August 2005.
- Z. Yang, K. Tang, and X. Yao, “Large scale evolutionary optimization using cooperative coevolution,” Information Sciences, vol. 178, no. 15, pp. 2985–2999, 2008.
- X. Li and X. Yao, “Tackling high dimensional nonseparable optimization problems by cooperatively coevolving particle swarms,” in Proceedings of the Congress on Evolutionary Computation, CEC '9, pp. 1546–1553, IEEE Computational Intelligence Magazine, Trondheim, Norway, May 2009.
- M. N. Omidvar, X. Li, Z. Yang, and X. Yao, “Cooperative co-evolution for large scale optimization through more frequent random grouping,” in Proceedings of the 2010 6th IEEE World Congress on Computational Intelligence, WCCI 2010 - 2010 IEEE Congress on Evolutionary Computation, CEC 2010, Spain, July 2010.
- T. Ray and X. Yao, “A cooperative coevolutionary algorithm with correlation based adaptive variable partitioning,” in Proceedings of the 2009 IEEE Congress on Evolutionary Computation (CEC), pp. 983–989, IEEE, Trondheim, Norway, May 2009.
- W. Chen, T. Weise, Z. Yang, and K. Tang, “Large-scale global optimization using cooperative coevolution with variable interaction learning,” in Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics): Preface, vol. 6239, pp. 300–309, 2010.
- M. N. Omidvar, X. Li, Y. Mei, and X. Yao, “Cooperative co-evolution with differential grouping for large scale optimization,” IEEE Transactions on Evolutionary Computation, vol. 18, no. 3, pp. 378–393, 2014.
- M. N. Omidvar, X. Li, and X. Yao, “Smart use of computational resources based on contribution for cooperative co-evolutionary algorithms,” in Proceedings of the 13th Annual Genetic and Evolutionary Computation Conference, GECCO'11, pp. 1115–1122, IEEE, Ireland, July 2011.
- R. Storn, “On the usage of differential evolution for function optimization,” in Proceedings of the Biennial Conference of the North American Fuzzy Information Processing Society (NAFIPS '96), pp. 519–523, June 1996.
- X. Li and X. Yao, “Cooperatively coevolving particle swarms for large scale optimization,” IEEE Transactions on Evolutionary Computation, vol. 16, no. 2, pp. 210–224, 2012.
- K. Weicker and N. Weicker, “On the improvement of coevolutionary optimizers by learning variable interdependencies,” in Proceedings of the 1999 Congress on Evolutionary Computation, CEC 1999, pp. 1627–1632, IEEE, July 1999.
- E. Sayed, D. Essam, and R. Sarker, “Using hybrid dependency identification with a memetic algorithm for large scale optimization problems,” in Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics): Preface, vol. 7673, pp. 168–177, 2012.
- Y. Liu, X. Yao, Q. Zhao, and T. Higuchi, “Scaling up fast evolutionary programming with cooperative coevolution,” in Proceedings of the 2001 Congress on Evolutionary Computation, pp. 1101–1108, IEEE, Seoul, South Korea, 2001.
- E. Sayed, D. Essam, and R. Sarker, “Dependency identification technique for large scale optimization problems,” in Proceedings of the 2012 IEEE Congress on Evolutionary Computation, CEC 2012, Australia, June 2012.
- Y. Ren and Y. Wu, “An efficient algorithm for high-dimensional function optimization,” Soft Computing, vol. 17, no. 6, pp. 995–1004, 2013.
- J. Liu and K. Tang, “Scaling up covariance matrix adaptation evolution strategy using cooperative coevolution,” in Proceedings of the International Conference on Intelligent Data Engineering and Automated Learning – IDEAL 2013, vol. 8206 of Lecture Notes in Computer Science, pp. 350–357, Springer, Berlin, Germany, 2013.
- Z. Yang, K. Tang, and X. Yao, “Multilevel cooperative coevolution for large scale optimization,” in Proceedings of the IEEE Congress on Evolutionary Computation (CEC '08), pp. 1663–1670, June 2008.
- M. N. Omidvar, Y. Mei, and X. Li, “Effective decomposition of large-scale separable continuous functions for cooperative co-evolutionary algorithms,” in Proceedings of the 2014 IEEE Congress on Evolutionary Computation, CEC 2014, pp. 1305–1312, China, July 2014.
- S. Mahdavi, E. M. Shiri, and S. Rahnamayan, “Cooperative Co-evolution with a new decomposition method for large-scale optimization,” in Proceedings of the 2014 IEEE Congress on Evolutionary Computation (CEC), 2014.
- B. Kazimipour, M. N. Omidvar, X. Li, and A. K. Qin, “A sensitivity analysis of contribution-based cooperative co-evolutionary algorithms,” in Proceedings of the IEEE Congress on Evolutionary Computation, CEC 2015, pp. 417–424, Japan, May 2015.
- Z. Cao, L. Wang, Y. Shi et al., “An effective cooperative coevolution framework integrating global and local search for large scale optimization problems,” in Proceedings of the IEEE Congress on Evolutionary Computation, CEC 2015, pp. 1986–1993, Japan, May 2015.
- X. Peng, K. Liu, and Y. Jin, “A dynamic optimization approach to the design of cooperative co-evolutionary algorithms,” Knowledge-Based Systems, vol. 109, pp. 174–186, 2016.
- H. K. Singh and T. Ray, “Divide and conquer in coevolution: a difficult balancing act,” in Agent-Based Evolutionary Search, vol. 5 of Adaptation, Learning, and Optimization, pp. 117–138, Springer, Berlin, Germany, 2010.
- X. Peng and Y. Wu, “Large-scale cooperative co-evolution using niching-based multi-modal optimization and adaptive fast clustering,” Swarm & Evolutionary Computation, vol. 35, 2017.
- M. Yang, M. N. Omidvar, C. Li et al., “Efficient resource allocation in cooperative co-evolution for large-scale global optimization,” IEEE Transactions on Evolutionary Computation, vol. 21, no. 4, pp. 493–505, 2017.
- M. N. Omidvar, X. Li, and X. Yao, “Cooperative co-evolution with delta grouping for large scale non-separable function optimization,” in Proceedings of the 2010 IEEE Congress on Evolutionary Computation (CEC 2010), pp. 1–8, Barcelona, Spain, July 2010.
- X. Peng and Y. Wu, “Enhancing cooperative coevolution with selective multiple populations for large-scale global optimization,” Complexity, vol. 2018, Article ID 9267054, 15 pages, 2018.
- S.-T. Hsieh, T.-Y. Sun, C.-C. Liu, and S.-J. Tsai, “Solving large scale global optimization using improved particle swarm optimizer,” in Proceedings of the 2008 IEEE Congress on Evolutionary Computation, CEC 2008, pp. 1777–1784, China, June 2008.
- T. Hatanaka, T. Korenaga, N. Kondo et al., Search Performance Improvement for PSO in High Dimensional Space, InTech, 2009.
- J. Fan, J. Wang, and M. Han, “Cooperative coevolution for large-scale optimization based on kernel fuzzy clustering and variable trust region methods,” IEEE Transactions on Fuzzy Systems, vol. 22, no. 4, pp. 829–839, 2014.
- A.-R. Hedar and A. F. Ali, “Genetic algorithm with population partitioning and space reduction for high dimensional problems,” in Proceedings of the 2009 International Conference on Computer Engineering and Systems, ICCES'09, pp. 151–156, Egypt, December 2009.
- J. Q. Zhang and A. C. Sanderson, “JADE: adaptive differential evolution with optional external archive,” IEEE Transactions on Evolutionary Computation, vol. 13, no. 5, pp. 945–958, 2009.
- L. M. Hvattum and F. Glover, “Finding local optima of high-dimensional functions using direct search methods,” European Journal of Operational Research, vol. 195, no. 1, pp. 31–45, 2009.
- C. Liu and B. Li, “Memetic algorithm with adaptive local search depth for large scale global optimization,” in Proceedings of the 2014 IEEE Congress on Evolutionary Computation, CEC 2014, pp. 82–88, China, July 2014.
- S. Saremi, S. Z. Mirjalili, and S. M. Mirjalili, “Evolutionary population dynamics and grey wolf optimizer,” Neural Computing and Applications, vol. 26, no. 5, pp. 1257–1263, 2015.
- A. Zhu, C. Xu, Z. Li, J. Wu, and Z. Liu, “Hybridizing grey wolf optimization with differential evolution for global optimization and test scheduling for 3D stacked SoC,” Journal of Systems Engineering and Electronics, vol. 26, no. 2, pp. 317–328, 2015.
- V. Chahar and D. Kumar, “An astrophysics-inspired grey wolf algorithm for numerical optimization and its application to engineering design problems,” Advances in Engineering Software, vol. 112, pp. 231–254, 2017.
- D. Guha, P. K. Roy, and S. Banerjee, “Load frequency control of large scale power system using quasi-oppositional grey wolf optimization algorithm,” Engineering Science and Technology, an International Journal, vol. 19, no. 4, pp. 1693–1713, 2016.
- C. Lu, S. Xiao, X. Li, and L. Gao, “An effective multi-objective discrete grey wolf optimizer for a real-world scheduling problem in welding production,” Advances in Engineering Software, vol. 99, pp. 161–176, 2016.
- E. Emary, H. M. Zawbaa, and A. E. Hassanien, “Binary grey wolf optimization approaches for feature selection,” Neurocomputing, vol. 172, pp. 371–381, 2016.
- W. Long, J. Jiao, X. Liang, and M. Tang, “Inspired grey wolf optimizer for solving large-scale function optimization problems,” Applied Mathematical Modelling: Simulation and Computation for Engineering and Environmental Systems, vol. 60, pp. 112–126, 2018.
- S. Gupta and K. Deep, “Performance of grey wolf optimizer on large scale problems,” in Proceedings of the Mathematical Sciences and Its Applications, American Institute of Physics Conference Series, Uttar Pradesh, India, 2017.
- R. L. Haupt and S. E. Haupt, “Practical genetic algorithms,” Journal of the American Statistical Association, vol. 100, no. 100, p. 253, 2005.
- H. R. Tizhoosh, “Opposition-based learning: a new scheme for machine intelligence,” in Proceedings of the International Conference on Computational Intelligence for Modelling, Control and Automation, CIMCA and International Conference on Intelligent Agents, Web Technologies and Internet Commerce (IAWTIC '05), pp. 695–701, Vienna, Austria, November 2005.
- H. Wang, Z. Wu, S. Rahnamayan, Y. Liu, and M. Ventresca, “Enhancing particle swarm optimization using generalized opposition-based learning,” Information Sciences, vol. 181, no. 20, pp. 4699–4714, 2011.
- H. Wang, S. Rahnamayan, and Z. Wu, “Parallel differential evolution with self-adapting control parameters and generalized opposition-based learning for solving high-dimensional optimization problems,” Journal of Parallel and Distributed Computing, vol. 73, no. 1, pp. 62–73, 2013.
- X. Yao, Y. Liu, and G. Lin, “Evolutionary programming made faster,” IEEE Transactions on Evolutionary Computation, vol. 3, no. 2, pp. 82–102, 1999.
- A. P. Engelbrecht, Fundamentals of Computational Swarm Intelligence, John Wiley & Sons, 2006.
- S. Mirjalili and A. Lewis, “The whale optimization algorithm,” Advances in Engineering Software, vol. 95, pp. 51–67, 2016.
- S. Mirjalili, A. H. Gandomi, S. Z. Mirjalili, S. Saremi, H. Faris, and S. M. Mirjalili, “Salp Swarm Algorithm: A bio-inspired optimizer for engineering design problems,” Advances in Engineering Software, vol. 114, pp. 163–191, 2017.
- S. Mirjalili, “The ant lion optimizer,” Advances in Engineering Software, vol. 83, pp. 80–98, 2015.
- Q. Gu, S. Jiang, M. Lian, and C. Lu, “Health and safety situation awareness model and emergency management based on multi-sensor signal fusion,” IEEE Access, vol. 7, pp. 958–968, 2019.
Copyright © 2019 Qinghua Gu et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.