Solving Engineering and Science Problems Using Complex Bio-inspired Computation ApproachesView this Special Issue
Review Article | Open Access
Hanfeng Zhou, Zewei Ding, Hongxin Peng, Zitao Tang, Guoxi Liang, Huiling Chen, Chao Ma, Mingjing Wang, "An Improved Grasshopper Optimizer for Global Tasks", Complexity, vol. 2020, Article ID 4873501, 23 pages, 2020. https://doi.org/10.1155/2020/4873501
An Improved Grasshopper Optimizer for Global Tasks
The grasshopper optimization algorithm (GOA) is a metaheuristic algorithm that mathematically models and simulates the behavior of the grasshopper swarm. Based on its flexible, adaptive search system, the innovative algorithm has an excellent potential to resolve optimization problems. This paper introduces an enhanced GOA, which overcomes the deficiencies in convergence speed and precision of the initial GOA. The improved algorithm is named MOLGOA, which combines various optimization strategies. Firstly, a probabilistic mutation mechanism is introduced into the basic GOA, which makes full use of the strong searchability of Cauchy mutation and the diversity of genetic mutation. Then, the effective factors of grasshopper swarm are strengthened by an orthogonal learning mechanism to improve the convergence speed of the algorithm. Moreover, the application of probability in this paper greatly balances the advantages of each strategy and improves the comprehensive ability of the original GOA. Note that several representative benchmark functions are used to evaluate and validate the proposed MOLGOA. Experimental results demonstrate the superiority of MOLGOA over other well-known methods both on the unconstrained problems and constrained engineering design problems.
The swarm intelligence technique is used to simulate the behavior of different groups that cooperatively search for food in nature. By learning from each other, each member of these groups continually shifts the direction of searching [1–8]. Compared with the traditional optimization methods , they show unique advantages in optimizing practical problems [10–17]. Some typical algorithms and as particle swarm optimization (PSO) , differential evolution (DE) , artificial bee colony optimization (ABC) , firefly algorithm (FA) , fruit fly optimization algorithm (FOA) , Harris hawks optimization (HHO) , slime mould algorithm (SMA) , and grasshopper optimization algorithm (GOA) . Because of its unique advantages, GOA has outstanding performance in solving practical problems and has been widely concerned in recent years.
Since its introduction, GOA has been applied to tackling many problems. Aljarah et al.  introduced GOA to find the variables in the design of the support vector machine (SVM) and to evaluate the feature functions of the ideal subsets. This technique minimizes the amount of chosen characteristics and is superior to other techniques in classification precision. The optimized GOA was introduced to the electrical field by Barman et al. . A regional hybrid short-term load forecast (STLF) model based on GOA with SVM was suggested to assess the relevant parameters in order to enhance precision. Hamad et al.  applied enhanced GOA combined with SVM for automatic detection of EEG epilepsy. The results showed that this method has better classification accuracy. Ibrahim et al.  combined SVM with improved GOA for feature selection and applied the method to Iraqi cancer patients and biomedical datasets (California Irvine datasets) to achieve high-performance accuracy in real-world datasets. In order to accelerate its global convergence rate, Arora and Anand.  proposed improved GOA by using chaos theory to balance the optimization process of GOA. Saxena et al.  used chaotic mapping to enhance GOA and applied it to the parameter estimation of three bar truss layout and FM acoustic synthesis. El-Fergany  took the modeling of the proton exchange membrane fuel cell reactor as an example and the findings highlighted the feasibility and validity of the GOA-based program in steady-state and dynamic simulation. Fathy  used the objective function to maximize the set of slaves and matched GOA with the complete cross-connection (TCT), the Su Do Ku connection, and the GA set up to boost the output capacity of the photovoltaic array. Ewees et al.  proposed the modified GOA by utilizing a learning strategy, named opposition-based learning strategy (OBLGOA). Heidari et al.  used hybrid random training to improve GOA in terms of avoiding falling into the local optimal solution and balancing the trends of exploration and exploitation and applied this enhanced method to the medical field. Liu et al.  suggested a detailed approach mixing linear weighted sum (LWS) with GOA to address the energy management issue, improve exhaustive energy efficiency, and ensure regional coordination and optimization. Luo et al.  put forward an enhanced GOA system combining three approaches to achieve a more suitable development-exploration equilibrium. Mafarja et al.  chose the enhanced binary GOA to subject it to the selection of the ideal subset of features and that of the wrapper-based classification field, which made the algorithm’s output superior. Mirjalili et al.  introduced a selection strategy into GOA to make it have the ability to deal with multiobjective issues, which made the algorithm archive a good maintenance in terms of high convergence and adaptability. Tharwat et al.  integrated an external archive with GOA to preserve the optimal solution; then, this archive was used in multiobjective search space to define the social behavior of GOA and apply it to the field of multiobjective problems. Wu et al.  combined comprehensive empirical mode decomposition (EMD) with extreme learning machines and GOA and applied them to STLF in power management. The improvement in the algorithm can find the most appropriate parameters of the extreme learning machine. Wu et al.  improved the GOA by developing a natural selection approach, a democratic selection-making mechanism, and a vibrant 1/5-based feedback mechanism. Zakeri and Hokmabadi  used statistical measures to complement the iteration process that replaced repetitive features with the most promising ones, and they were applied to the field of feature selection in a final. Zhang et al.  suggested a GOA-based parameter-adaptive variational mode decomposition (VMD) technique for analyzing vibration signals of rotating machinery. Zhao et al.  merged the technique of decomposition of Beveridge-Nelson (B-N) with the least squares support vector machine (LSSVM) and GOA and applied it to the prediction model of wind power.
This paper introduces two approaches to GOA and proposes an enhanced version called MOLGOA. First, taking into account the benefits and disadvantages of Cauchy mutation and genetic mutation, we incorporated the probabilistic combination mutation system into GOA to enhance the algorithm’s worldwide search capability. Then we applied the orthogonal learning mechnism (OL) to GOA, which significantly enhanced the algorithm’s convergence velocity. In order to check the effectiveness of the enhanced MOLGOA, the algorithm was evaluated on some representative benchmark functions, including single-mode, multimode, and fixed-dimensional models, and the test results were compared with other competitors including mature metaheuristic algorithms and variations of several sophisticated algorithms. Experiments show that the mechanism proposed in this paper effectively alleviates the problems of local optimization and slow convergence of the original GOA and solves the balance the diversification and intensification.
The remainder of this paper is structured as follows. The original GOA is presented in Section 2. Section 3 shows the details of the enhanced MOLGOA. The fourth section presents and discusses the experimental outcomes. MOLGOA is defined for engineering issues in Section 5. Finally, conclusions and future works are summarized.
2. An Overview of GOA
GOA was proposed by Saremi et al.  in 2017, which simulates the population behavior of grasshopper. Grasshoppers are pests whose life cycle consists of eggs, nymphs, and adults that affect crop production and agriculture. GOA simulates grasshopper’s behavior in nymph and adult stages. In the nymph stage, grasshopper’s primary motion features are jumping and moving in the manner of rolling cylinders (tiny steps, slow motion) and eat crops discovered along with the manner, while in the phase of adulthood, grasshoppers migrate in groups over lengthy distances (quick steps).
GOA simulates grasshopper population behavior according to the following formula:where Xi is the position of the ith grasshopper and Si means social interaction. Gi denotes the gravity of the i grasshopper and Ai denotes the gravity wind advection of the i grasshopper.
Firstly, the S component in equation (1) is calculated according to the following equation:
dij represents the distance between grasshopper i and grasshopper j, which is defined as , and represents the distance unit vector from grasshopper i to grasshopper j.
The population power function s(r) is described as follows:where f reflects the intensity of the appeal and the duration of the attraction.
Secondly, the G component in equation (1) is obtained aswhere denotes the gravitational constant, and is a unit vector pointing to the origin.
Thirdly, the component A in equation (3) is determined by the following equation:where u and represent a unit vector of constant drift and wind direction, respectively. Then combining S, G, and A into one formula, the expression is changed as follows:where N represents the number of grasshoppers. The coefficient c of each iteration is determined according to the following equation:where and represent the upper bound and lower bound of c, respectively, l is the current iteration, and L represents the maximum iteration number.
3. Improved MOLGOA Method
The improved MOLGOA has two effective strategies. Firstly, the orthogonal mechanism is successfully integrated with GOA to accelerate the convergence speed of it. Then, for the advantages of Gaussian mutation and nonlinear programming mutation, these two mutation strategies are adopted in the improved MOLGOA.
3.1. Orthogonal Learning Mechanism
The orthogonal learning mechanism has been proved to be effective in lifting the optimization capability of the intelligent algorithms [46–53]. The orthogonal table is obtained by listing out the horizontal, vertical, and mixed orthogonal table selected in the orthogonal test design. Based on the in-depth study of mathematical combination theory, a set of test design table with complete rules is constructed by using Latin letters. The symbol of the orthogonal table is : L—The sign of an orthogonal table M—Number of rows in the orthogonal table (number of tests, number of test schemes) Q—Number of digits in an orthogonal table (level of factors) F—Number of columns in orthogonal tables (number of experimental factors)
According to the data structure of the orthogonal table, it can be seen that its distribution is a “crisscross” table, where “vertical” refers to the horizontal distribution of the factors selected in the experiment, and “horizontal” refers to the number of factors examined in the implementation. The total number of “longitudinal” is the exact amount of experimental schemes of the orthogonal test. As an orthogonal table of row F and column Q, its characteristics are mainly described as in each column; different numbers appear the same number of times and the numbers in any two columns are arranged in a complete and balanced manner. This fully reflects the two advantages of the orthogonal table, namely, “uniform dispersion, neatly comparable”. In general, each level of each factor touches each level of the other factors, and that’s orthogonal. Based on the orthogonal combination and factor analysis, the highest combination can be readily discovered, which can be proved in the literature .
The orthogonal experimental structure is efficient for solving the  multi-factor and multi-level experimental issues. By arranging a few tests with the orthogonal table, it can discover the finest way for implementation circumstances like optimization. is used for orthogonal table where F and Q represented as factor and levels respectively. Similarly, L and M denoted as Latin square and number of horizontal combinations, respectively. For the experimental problems of F factor and Q level, there are groups of experiments if the comprehensive combination experiment is carried out. However, it is impossible to do all groups of experiments when Q and F are large. So, the orthogonal table is applied to arrange the orthogonal experiment, which makes only M combinations needed for the experiment. Certainly, M is usually much less than . We take orthogonal table as an example, which is shown in Table 1. For problems of 3 factors and 3 levels, only 9 experiments are needed to conduct orthogonal experiments according to orthogonal table , but 33 = 27 experiments are needed for comprehensive combined experiments in total. It can be seen that orthogonal experimental design greatly reduces the number of experiments, and the greater the factors and level, the more obvious advantages of this method.
For convenience, mark where the ith combination of the level of j is arranged as . The jth column of the orthogonal table illustrated as , if , then is called the basic column; the other columns are called the nonbasic columns. A previous study  demonstrated an algorithm for generating the orthogonal table that first generates fundamental columns and then generates nonbasic columns. As described in Table 2, Q is the primary column and , J satisfies the formula .
Through the orthogonal table, we can find the optimal combination between different levels of various factors with fewer test times. By comparing the characteristics of uniform dispersion and the neatness of the orthogonal table, the algorithm is not only implemented on a smaller scale but also accelerates the convergence rate of itself.
3.2. Genetic Mutation
Genetic algorithm  is an evolutionary algorithm whose fundamental concept is to imitate the evolutionary law of the biological world’s “survival of the fittest in natural selection”. Genetic algorithms encode issue parameters as chromosomes and use iterative selection, crossover, mutation, and other procedures to exchange chromosome data in the population in order to produce chromosomes in line with optimization goals.
Nonlinear programming is a new discipline formed in the 1950s that is used under a set of constraints of equality or inequality to study the extreme problem of the n-element real function.
The constraint conditions arewhere, , , , , and are vectors. and are matrices. and return functions of vectors. , , and are nonlinear functions.
The mutation procedure appears in the process of using a genetic algorithm to fix this issue, whose primary aim is to maintain the diversity of the population. This mutation procedure chooses a person from a population randomly and then chooses a point in the person to mutate to create a better person. The mutation operation of the gene in the ith individual is formulated as follows:where and are the maximum and minimum of , respectively:
In equation (10), is a random number, represents the current iteration number, is the maximum evolution number, and is a randomly generated value between 0 and 1.
By referring to the GA-based mutation system for solving nonlinear programming problems, new mutant individuals can be acquired according to a certain variety of probability; the potential individual can be acquired according to a certain variety of probability, which enhances the algorithm’s search capability and retains population diversity.
3.3. Cauchy Mutation
The Cauchy mutation is used to enhance the search capabilities of algorithms in the context of searching and also to avoid the grasshopper population from falling into local optimum. The cauchy mutation is to produce a set of random variables that are consistent with the Cauchy distribution to change the population’s position to increase population searchability. The graph of Cauchy density function  is quite similar to that of the Gaussian variation density function. The difference between the two is that Cauchy density function slides slowly on both sides of the Y-axis. Moreover, the Cauchy distribution with this feature means that the Cauchy distribution can produce a large disturbance to avoid falling into local optimum and have a stronger global search ability.
The Cauchy mutation is formulated as the following formula:
The one-dimensional density function of Cauchy mutation is determined as
So when , , the standard form of the one-dimensional Cauchy density function at the center of the origin is obtained as follows:
Correspondingly, the standard Cauchy distribution function is illustrated as follows:
The advantage of Cauchy mountain is to generate high-probability disturbance to enhance global search capability. The Cauchy mutation is widely used in various optimization algorithms, such as PSO , which uses a Cauchy mutation to mutate the optimal particles obtained and achieved the optimal global solution. Another instance is the krill herd optimization algorithm (KHA)  that utilizes the Cauchy mutation to achieve better alternatives by optimizing the present situation.
3.4. Combined Mutation Strategy
In this subsection, the technique of variability in probability choice is implemented to enhance GOA’s search precision and worldwide search capability. Because of the benefits of Cauchy mutation in improving worldwide searchability and the features of potential individuals produced by genetic algorithm, Cauchy mutation is first performed, and then the iteration genetic mutation is feasible. The steps are given as follows: Step 1: Judge the times of the present iteration. If it is not a multiple of 5, create a one-dimensional row vector that is compatible with equation (11) and perform Cauchy mutation on each individual’s present place in the population. At the same moment, determine whether the position is superior to the original position after Cauchy mutation. If so, there will be a Cauchy mutation; otherwise, the original position will be retained. If the current number of iterations is a multiple of 5, then jump to Step 2 directly. Step 2: Generate a number n randomly (0 < n < 1) and continue to Step 3 to the current genetic mutation limit. Step 3: Equation (9) carries out genetic mutation on each individual’s present place in the population. (Note that the mutated position does not need to be compared with the original position to join the following iteration.) The proposed MOLGOA steps are illustrated in Algorithm 1, and the procedure of MOLGOA is presented in the form of a flowchart given in Figure 1.
In Figure 1, is the current number of iterations, is the maximum number of iterations, is the population at the random initial position, is the population by carrying out mutation operation, is the population after Cauchy mutation, and is the best fitness.
In this paper, the time complexity of this algorithm mainly depends on the algorithm (E) and the total number of problems (d). Through analysis, it can be concluded that the overall time complexity of this algorithm is O (MOLGOA) = O (initialization of the population position) + O (calculation of the individual comfort) + O (choice of the best comfort level) + E × (O (update of the individual location) + O (choice of the variation) + O (orthogonal conduction and Choice of the best comfort level) + O (calculation of the individual comfort and choice of the best comfort level).
The time complexity of initializing population location is O (n × d), the time complexity of the calculation of individual comfort is O (n), the time complexity of the choice of the optimal comfort is O (n), the time complexity of the update of the individual position to O (n × n × d), the time complexity of the choice of the variation is O (n), the time complexity of conduction of orthogonal learning and choice of the best comfort level is O(n × (d + n × d × d)), and the time complexity of calculating individual comfort and the choice of the best comfort level is O (n).
Therefore, the time complexity of the algorithm proposed in this paper is O (MOLGOA) = O (n × d) + O (n) + O (n) + E × (O(n × n × d) + O(n) + O(n × (d + n × d × d)) +On)) = O(n × d) + 2O(n) + E × (O(n × n × d) + O(n × (d + n × d × d)) + 2O(n)).
4. Experimental Results and Discussions
In this section, various experiments on optimization benchmark problems are implemented, and the results are discussed in detail.
4.1. Comparative Results on CEC2017 Problems
To check the optimizer’s efficiency, 30 well-known benchmark functions are chosen. Functions are listed in Table 3, which can be split into three classifications: F1–F3 are single combined transports, F4–F10 are multimodal transports, F11–F20 are blended functions, and F21–F30 are combined functions. Since they have only one worldwide best and no local optimum (LO), monomodal features can be implemented to benchmark development capacities. Compared with single-mode functions, multimode functions have a large amount of LO and can boost with volume, which can be used to test the algorithm’s capacity of exploration and that of jumping out of LO. Furthermore, the multimodal fixed-dimensional transport functions only have one ideal global solution but a large amount of LOs, as well as multimodal transport functions. However, since the former’s solution space is tiny, it is necessary to reduce the search agent’s adaptive step size. Therefore, in order to comprehensively assess the algorithm’s efficiency and tackle with various kinds of issues, thirty benchmark features are used to evaluate the algorithms.
As mentioned above, the solution space is tiny; therefore, the search agent’s adaptive step size is low. Thus, to investigate the algorithm’s efficiency and for solving various types of problems, and evaluate the optimization algorithm’s performance comprehensively, thirty functions are adopted for the experimental tests.
For the fairness of the trial, all tests were carried out by 30 search agents for 1,000 iterations under the same conditions. Furthermore, it should be observed that to mitigate the impacts of randomness, each benchmark feature is performed 30 times separately.
4.2. The Impact of Orthogonal Learning and Combined Mutation
As can be seen from the above sections of this paper, two strategies, namely, combinatorial mutation and orthogonal mechanism, have been implemented into the initial GOA. We created three distinct GOA algorithms to study the impact of each mechanism and mixture, as shown in Table 4. “M” and “OL” mean “combinatorial mutation” and “orthogonal mechanism,” respectively. In Table 4, “1” implies that this operation is integrated into GOA, and “0” implies that there is no corresponding operation. For instance, OLGOA implies that GOA combines the orthogonal mechanism, excluding combinatorial mutations.
The efficiency of various enhanced GOA variants was evaluated based on these functions, as shown in Table 4. Because nonparametric statistical tests were usually adopted in optimization algorithms comparison, the Wilcoxon sign rank test  and the Friedman test  have been conducted. Table 5 shows the statistical outcomes of different enhanced GOA variants. First, to study the statistical significance distinction between pairing algorithms, the nonparametric Wilcoxon sign rank test with a meaning rate of 5 percent was used. It is shown in Table 5, the values of MOLGOA and other competitors in 30 functions (F1–F30) are given in rows 2 through 5. The symbols of “+,” “−,” and “ = ” show that MOLGOA is higher than, lower than, and equal to other algorithms, respectively. MOLGOA showed statistically significant distinctions in the situation of “+/−/ = ” compared to these techniques of comparison. MOLGOA is usually better than OLGOA, MGOA, and GOA among the 30 functions, as you can see from the figure, so MOLGOA is the best performing variant. In addition, to further explore the distinctions between different techniques, in order to evaluate the average classification value of all these algorithms, the Friedman test is used in the experiments. Obviously, the combination of the two approaches has the smallest value according to these sorting outcomes, which verifies that MOLGOA performs better than other combinations in solving these test functions. Lastly, based on the above assessment, MOLGOA is chosen as GOA’s best technique of enhancement.
4.3. Comparison with Other Well-Established Methods
In this section, the performance of the MOLGOA will be validated by comparing with the original GOA and some more successful heuristic methods such as gray wolf optimizer (GWO), whale optimization algorithm (WOA) , sine cosine algorithm (SCA) , differential evolution algorithm (DE), dragonfly algorithm (DA), bat algorithm (BA), moth-flame optimization algorithm (MFO), particle swarm optimization (PSO), Harris hawks optimization (HHO) , and GOA . The dimension of each algorithm to be tested is set to 30. Table 6 shows the detailed comparison results.
Table 6 shows that MOLGOA considerably increases the initial GOA’s performance. Furthermore, in Table 6, all values indicate that MOLGOA has the lowest deviation compared to other techniques. MOLGOA has better stability compared to other techniques and is able to locate the ideal solution in a narrower spectrum. As shown in Table 7, it can be seen from the “+/−/=” line that MOLGOA accomplished the outstanding outcomes. The findings indicate that MOLGOA has evident benefits and statistical significance in different kinds of functions compared to other rivals. We performed the Friedman test on the median ranking value shown in Table 8 further to investigate the distinctions between MOLGOA and the other rivals. From the outcomes, it can be evidently seen that MOLGOA has reached the highest quality of the solutions with the average ranking value of 1.16667.
Figure 2 demonstrates the convergence characteristics of MOLGOA and other counterparts on various 30-dimensional benchmark issues. As shown, the MOLGOA has the greatest convergence precision and efficiency when it comes to F3, F14, F18, F22, and F26 issues, while other algorithms fall prematurely into local optimization. However, they all have comparatively decent convergence precision in F1, F5, and F10. In sum, MOLGOA can achieve better convergence curve and shows a quicker convergence rate and can find the best solution to the issue compared to other rivals. From the above assessment, it can be seen that MOLGOA displays great global search capability and high search precision for low-dimensional tasks as well as high-dimensional tasks, so we can conclude that MOLGOA is superior to those of algorithms in this domain.
4.4. Comparison with the State-of-The-Art Algorithms
To further evaluate the performance of MOLGOA, this impressive method was compared with other sophisticated algorithms in this chapter, including improved GOA (IGOA) , CSSA , ESSA , CGSCA , m_SCA , LWOA , BWOA , CCMWOA , CLPSO , and GL25 . The detailed illustration of the comparison functions and the parameters involved in the experiment are listed in Table 3.
In this experiment, Table 9 gives the mean and standard deviation values achieved by these algorithms, while the Wilcoxon test demonstrates the final outcomes of the comparison of distinct algorithms. Table 10 demonstrates the outcomes of the Friedman test, which can demonstrate the benefits of each algorithm more obviously.
According to the information shown in Table 9, it can be found that MOLGOA is superior to other sophisticated algorithms in most cases. As shown in Table 11, we can find that MOLGOA has strong competitiveness among the selected functions in the Wilcoxon test, and the last row of Table 10 shows that MOLGOA performs better in most problems than other advanced methods. Table 11 lists the final ranking outcomes acquired through the Friedman test by distinct algorithms. The convergence rate of MOLGOA is presented in Figure 3; it can be seen that the convergence speed of the proposed algorithm is the fastest among all the methods.
According to the convergence curves in Figure 3, it can be seen that MOLGOA has the greatest convergence precision in all these functions. In F1, although MOLGOA’s convergence rate is slow at an early point, in the later phase, MOLGOA can jump from the location optimum and have high convergence precision. As shown, it is evident that MOLGOA has better convergence precision and quicker convergence rate on F2, F6, F14, F18, F22, and F26 functions.
MOLGOA has improved efficiency in the single modal functions, multimodal functions, and fixed-dimensional multimodal functions compared to other advanced methods. This is primarily because MOLGOA adopts the orthogonal mechanism and the joint mutation mechanism, which significantly enhances the algorithm’s global search capability, while the orthogonal mechanism offers efficient variables with the improvement feature. Overall, the benchmark issues can be efficiently solved by MOLGOA.
5. MOLGOA for the Engineering Benchmarks
MOLGOA is implemented in this chapter to deal with two engineering benchmark issues, namely, pressure vessel (PV) design problem and tension-compression string (TCS) design problem. The flowchart of MOLGOA for constraint optimization is shown in Figure 4.
5.1. PV Problem
This mathematical modeling is aimed at minimizing the total cost of the cylindrical pressure vessel, which is closely related to material, forming, and welding. In this optimization task, the thickness of the shell, the thickness of the head, the inner radius, and the length of the cylindrical section, without considering the head, are the variables to be optimized. The formulation can be described as bellow:where .
MOLGOA solves this issue and is compared to initial algorithms such as GOA, IGOA, and others. Table 12 shows the comparison outcomes, which indicates that the searching capability of MOLGOA is superior to other techniques. The findings of the comparison indicate that MOLGOA suggested in this article can also fix this issue efficiently.
5.2. TCS Problem
TCS problem requires minimizing the weight of a tension-compression spring that is subject to minimum deflection, shear stress, surge frequency, outer diameter limits, and design factors. The TCS issue solving model is defined as follows:
We adopted MOLGOA to solve the TCS problem and compared the results with other studies in the literature [70–77]. Mathematical approaches play an important role in engineering problems. However, there are still some problems in the mathematical approaches, such as the difficulty in solving the generalized geometric programming, the nondifferentiability of the dual objective function, and the divergence of the solutions of the sequential programming formed by the union terms. The proposed algorithm has the advantages of strong randomness and fewer parameters, which can avoid the above problems. The results recorded in Table 13 indicated that OLGOA found the minimum optimal cost. Compared with other existing methods, the effectiveness of this method is verified.
To conclude, our studies on these two classical engineering design issues demonstrate that the suggested MOLGOA can well optimize the practical issues. The algorithm’s efficacy in addressing constraint issues is further confirmed. The reason why MOLGOA is superior to other techniques in constraint issues is that the powerful exploration capability of MOLGOA can efficiently help the GOA in coordinating the inclination for exploration.
The main reason why the proposed MOLGOA can be applied to engineering optimization problems is the solution strategy formed by the mutual promotion of GOA, orthogonal mechanism, and joint mutation mechanism. Firstly, the basic equation of GOA is used to construct the population to maintain the diversity of the population. Secondly, Cauchy mutation has a significant ability to break through new areas in the exploration space, which means that the exploration ability of the population may be more powerful. A genetic algorithm is another mechanism for exploring new regions in solving nonlinear programming. The joint mutation mechanism, combined with these two different exploration methods, can help GOA to achieve the best solution faster and improve the convergence performance of the proposed algorithm. The orthogonal mechanism in the algorithm can reduce the search scale and make GOA find the global optimal value faster. Although the proposed MOLGOA has advantages in solving optimization problems, it may require considerable computational cost in the addition of orthogonal steps in MOLGOA. Therefore, although the orthogonal mechanism is very effective in some applications, a new mechanism should be studied in future research to dynamically select the most reasonable Q and F, so that MOLGOA can take into account the computational cost and get the best solution, to reduce the problem of high computational cost. At this stage, we mainly conducted a large number of experiments, continuously adjusted to select reasonable values for Q and F. However, in the existing experiments and engineering applications, we can see that the introduction of optimization mechanism in MOLGOA can greatly enhance the performance of GOA. We hope this method can be applied to other methods in the optimization field.
6. Conclusions and Future Works
In the presented study, MOLGOA is developed to alleviate the limits of the original GOA, such as insufficient exploration ability and slow convergence speed. This approach mainly introduces the orthogonal learning mechanism to make MOLGOA converge faster. Moreover, a combinatorial mutation strategy is proposed to realize the complementary advantages of Cauchy mutation and genetic mutation. The numerical results of typical benchmark functions show that the proposed MOLGOA can effectively tackle different types of optimization tasks. For the constrained optimization issues, the simulation results also illustrate that MOLGOA can enhance the calculation results and have practical value when solving two engineering design problems.
Of course, as a relatively new algorithm, there are still many problems worthy of further study because of the tremendous potential of GOA. Firstly, classical metaheuristic algorithms can be integrated into GOA to achieve a better balance between exploration and exploitation and improve the global search ability of GOA. Then, how to use MOLGOA to solve multiobjective tasks and dynamic optimization is scheduled as the next work. Besides, considering the time and space complexity of the algorithm when solving relatively complex problems, for further improving computational capacity and efficiency of the algorithm, GPU parallel algorithm and multithread parallel processing technology can be adopted to explore more possibilities. Finally, the exploration of the proposed method to other promising problems will be also interesting such as the parameter tuning for deep learning models [78–80], feature selection problems [81–84], social evolution modelling , and video coding optimization .
The data involved in this study are all public data, which can be downloaded through public channels.
Conflicts of Interest
The authors declare that there are no conflicts of interest regarding the publication of the article.
Guoxi Liang and Huiling Chen conceptualized the study. Hanfeng Zhou, Hongxin Peng, and Zitao Tang performed data curation. Guoxi Liang and Huiling Chen were responsible for funding acquisition. Hanfeng Zhou, Hongxin Peng, and Huiling Chen performed investigation. Hanfeng Zhou, Hongxin Peng, Zitao Tang, Chao Ma, Huiling Chen, Mingjing Wang were responsible for methodology. Hanfeng Zhou, Zewei Ding, Huiling Chen wrote the original draft. Guoxi Liang and Huiling Chen reviewed and edited the manuscript. Guoxi Liang and Huiling Chen contributed equally to this work.
This research was supported by the general research project of Zhejiang Provincial Education Department (Y201942618), Zhejiang Provincial Department of Education’s 2019 domestic visiting engineer of colleges and universities “School-Enterprise Cooperation Project” under Grant FG2019057, Zhejiang, China, and the National Natural Science Foundation of China (U1809209). The authors also appreciate the efforts of Ali Asghar Heidarihttp://www.aliasgharheidari.com for his support during the revision of this research.
- Y. Sun, G. G. Yen, and Z. Yi, “IGD indicator-based evolutionary algorithm for many-objective optimization problems,” IEEE Transactions on Evolutionary Computation, vol. 23, no. 2, pp. 173–187, 2019.
- Y. Sun, B. Xue, M. Zhang, and G. G. Yen, “Evolving deep convolutional neural networks for image classification,” IEEE Transactions on Evolutionary Computation, vol. 24, no. 2, pp. 394–407, 2020.
- Y. Sun, B. Xue, M. Zhang, G. G. Yen, and J. Lv, “Automatically designing CNN architectures using the genetic algorithm for image classification,” IEEE Transactions on Cybernetics, vol. 50, no. 9, pp. 3840–3854, 2020.
- H. Chen, Y. Xu, M. Wang, and X. Zhao, “A balanced whale optimization algorithm for constrained engineering design problems,” Applied Mathematical Modelling, vol. 71, pp. 45–59, 2019.
- J. Luo, H. Chen, A. A. Heidari, Y. Xu, Q. Zhang, and C. Li, “Multi-strategy boosted mutative whale-inspired optimization approaches,” Applied Mathematical Modelling, vol. 73, pp. 109–123, 2019.
- H. Yu, N. Zhao, P. Wang, H. Chen, and C. Li, “Chaos-enhanced synchronized bat optimizer,” Applied Mathematical Modelling, vol. 77, pp. 1201–1215, 2020.
- H. Chen, M. Wang, and X. Zhao, “A multi-strategy enhanced sine cosine algorithm for global optimization and constrained practical engineering problems,” Applied Mathematics and Computation, vol. 369, Article ID 124872, 2020.
- X. Zhang, Y. Xu, C. Yu et al., “Gaussian mutational chaotic fruit fly-built optimization and feature selection,” Expert Systems with Applications, vol. 141, Article ID 112976, 2020.
- X. Zhang, D. Wang, Z. Zhou, and Y. Ma, “Robust low-rank tensor recovery with rectification and alignment,” IEEE Transactions on Pattern Analysis and Machine Intelligence, p. 1, 2019.
- X. Zhao, D. Li, B. Yang, C. Ma, Y. Zhu, and H. Chen, “Feature selection based on improved ant colony optimization for online detection of foreign fiber in cotton,” Applied Soft Computing, vol. 24, pp. 585–596, 2014.
- M. Wang and H. Chen, “Chaotic multi-swarm whale optimizer boosted support vector machine for medical diagnosis,” Applied Soft Computing, vol. 88, p. 105946, 2020.
- X. Zhao, X. Zhang, Z. Cai et al., “Chaos enhanced grey wolf optimization wrapped ELM for diagnosis of paraquat-poisoned patients,” Computational Biology and Chemistry, vol. 78, pp. 481–490, 2019.
- X. Xu and H.-L. Chen, “Adaptive computational chemotaxis based on field in bacterial foraging optimization,” Soft Computing, vol. 18, no. 4, pp. 797–807, 2014.
- L. Shen, H. Chen, Z. Yu et al., “Evolving support vector machines using fruit fly optimization for medical data classification,” Knowledge-Based Systems, vol. 96, pp. 61–75, 2016.
- M. Wang, H. Chen, B. Yang et al., “Toward an optimal kernel extreme learning machine using a chaotic moth-flame optimization strategy with applications in medical diagnoses,” Neurocomputing, vol. 267, pp. 69–84, 2017.
- Y. Xu, H. Chen, J. Luo, Q. Zhang, S. Jiao, and X. Zhang, “Enhanced Moth-flame optimizer with mutation strategy for global optimization,” Information Sciences, vol. 492, pp. 181–203, 2019.
- H. Chen, Q. Zhang, J. Luo, Y. Xu, and X. Zhang, “An enhanced bacterial foraging optimization and its application for training kernel extreme learning machine,” Applied Soft Computing, vol. 86, p. 105884, 2020.
- J. Kennedy and R. Eberhart, “Particle swarm optimization,” in Proceedings of the ICNN’95—International Conference on Neural Networks, Perth, Australia, November 1995.
- R. Storn and K. Price, “Differential evolution—a simple and efficient heuristic for global optimization over continuous spaces,” Journal of Global Optimization, vol. 11, no. 4, pp. 341–359, 1997.
- D. Karaboga and B. Basturk, “A powerful and efficient algorithm for numerical function optimization: artificial bee colony (ABC) algorithm,” Journal of Global Optimization, vol. 39, no. 3, pp. 459–471, 2007.
- X. S. Yang, “Firefly algorithms for multimodal optimization,” in Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), pp. 169–178, Springer, Berlin, Germany, 2009.
- W.-T. Pan, “A new fruit fly optimization algorithm: taking the financial distress model as an example,” Knowledge-Based Systems, vol. 26, pp. 69–74, 2012.
- A. A. Heidari, S. Mirjalili, H. Faris, I. Aljarah, M. Mafarja, and H. Chen, “Harris hawks optimization: algorithm and applications,” Future Generation Computer Systems, vol. 97, pp. 849–872, 2019.
- S. Li, H. Chen, M. Wang, A. A. Heidari, and S. Mirjalili, “Slime mould algorithm: a new method for stochastic optimization,” Future Generation Computer Systems, vol. 111, pp. 300–323, 2020.
- S. Saremi, S. Mirjalili, and A. Lewis, “Grasshopper optimisation algorithm: theory and application,” Advances in Engineering Software, vol. 105, pp. 30–47, 2017.
- I. Aljarah, A. M. Al-Zoubi, H. Faris, M. A. Hassonah, S. Mirjalili, and H. Saadeh, “Simultaneous feature selection and support vector machine optimization using the grasshopper optimization algorithm,” Cognitive Computation, vol. 10, no. 3, pp. 478–495, 2018.
- M. Barman, N. B. Dev Choudhury, and S. Sutradhar, “A regional hybrid GOA-SVM model based on similar day approach for short-term load forecasting in Assam, India,” Energy, vol. 145, pp. 710–720, 2018.
- A. Hamad, E. H. Houssein, A. E. Hassanien, and A. A. Fahmy, “Hybrid grasshopper optimization algorithm and support vector machines for automatic seizure detection in EEG signals,” in The International Conference on Advanced Machine Learning Technologies and Applications (AMLTA2018), pp. 82–91, Springer, Berlin, Germany, 2018.
- H. T. Ibrahim, W. J. Mazher, O. N. Ucan, and O. Bayat, “A grasshopper optimizer approach for feature selection and optimizing SVM parameters utilizing real biomedical data sets,” Neural Computing and Applications, vol. 31, no. 10, pp. 5965–5974, 2019.
- S. Arora and P. Anand, “Chaotic grasshopper optimization algorithm for global optimization,” Neural Computing and Applications, vol. 31, no. 8, pp. 4385–4405, 2019.
- A. Saxena, S. Shekhawat, and R. Kumar, “Application and development of enhanced chaotic grasshopper optimization algorithms,” Modelling and Simulation in Engineering, vol. 2018, Article ID 4945157, 14 pages, 2018.
- A. A. El-Fergany, “Electrical characterisation of proton exchange membrane fuel cells stack using grasshopper optimiser,” IET Renewable Power Generation, vol. 12, no. 1, pp. 9–17, 2018.
- A. Fathy, “Recent meta-heuristic grasshopper optimization algorithm for optimal reconfiguration of partially shaded PV array,” Solar Energy, vol. 171, pp. 638–651, 2018.
- A. A. Ewees, M. Abd Elaziz, and E. H. Houssein, “Improved grasshopper optimization algorithm using opposition-based learning,” Expert Systems with Applications, vol. 112, pp. 156–172, 2018.
- A. A. Heidari, H. Faris, I. Aljarah, and S. Mirjalili, “An efficient hybrid multilayer perceptron neural network with grasshopper optimization,” Soft Computing, vol. 23, no. 17, pp. 7941–7958, 2019.
- J. Liu, A. Wang, Y. Qu, and W. Wang, “Coordinated operation of multi-integrated energy system based on linear weighted sum and grasshopper optimization algorithm,” IEEE Access, vol. 6, pp. 42186–42195, 2018.
- J. Luo, H. Chen, Q. Zhang, Y. Xu, H. Huang, and X. Zhao, “An improved grasshopper optimization algorithm with application to financial stress prediction,” Applied Mathematical Modelling, vol. 64, pp. 654–668, 2018.
- M. Mafarja, I. Aljarah, H. Faris, A. I. Hammouri, A. M. Al-Zoubi, and S. Mirjalili, “Binary grasshopper optimisation algorithm approaches for feature selection problems,” Expert Systems with Applications, vol. 117, pp. 267–286, 2019.
- S. Z. Mirjalili, S. Mirjalili, S. Saremi, H. Faris, and I. Aljarah, “Grasshopper optimization algorithm for multi-objective optimization problems,” Applied Intelligence, vol. 48, no. 4, pp. 805–820, 2018.
- A. Tharwat, E. H. Houssein, M. M. Ahmed, A. E. Hassanien, and T. Gabel, “MOGOA algorithm for constrained and unconstrained multi-objective optimization problems,” Applied Intelligence, vol. 48, no. 8, pp. 2268–2283, 2018.
- J. Wu, Z. Cui, Y. Chen, D. Kong, and Y.-G. Wang, “A new hybrid model to predict the electrical load in five states of Australia,” Energy, vol. 166, pp. 598–609, 2019.
- J. Wu, H. Wang, N. Li et al., “Distributed trajectory optimization for multiple solar-powered UAVs target tracking in urban environment by adaptive grasshopper optimization algorithm,” Aerospace Science and Technology, vol. 70, pp. 497–510, 2017.
- A. Zakeri and A. Hokmabadi, “Efficient feature selection method using real-valued grasshopper optimization algorithm,” Expert Systems with Applications, vol. 119, pp. 61–72, 2019.
- X. Zhang, Q. Miao, H. Zhang, and L. Wang, “A parameter-adaptive VMD method based on grasshopper optimization algorithm to analyze vibration signals from rotating machinery,” Mechanical Systems and Signal Processing, vol. 108, pp. 58–72, 2018.
- H. Zhao, H. Zhao, and S. Guo, “Short-term wind electric power forecasting using a novel multi-stage intelligent algorithm,” Sustainability (Switzerland), vol. 10, no. 3, p. 881, 2018.
- H. Zhang, A. A. Heidari, M. Wang, L. Zhang, H. Chen, and C. Li, “Orthogonal nelder-mead moth flame method for parameters identification of photovoltaic modules,” Energy Conversion and Management, vol. 211, Article ID 112764, 2020.
- H. Zhang, R. Li, Z. Cai et al., “Advanced orthogonal moth flame optimization with Broyden-Fletcher-Goldfarb-Shanno algorithm: framework and real-world problems,” Expert Systems with Applications, vol. 159, Article ID 113617, 2020.
- Y. Yang, H. Chen, S. Li, A. A. Heidari, and M. Wang, “Orthogonal learning harmonizing mutation-based fruit fly-inspired optimizers,” Applied Mathematical Modelling, vol. 86, pp. 368–383, 2020.
- W. Zhu, C. Ma, X. Zhao et al., “Evaluation of sino foreign cooperative education project using orthogonal sine cosine optimized kernel extreme learning machine,” IEEE Access, vol. 8, pp. 61107–61123, 2020.
- H. Chen, A. A. Heidari, X. Zhao, L. Zhang, and H. Chen, “Advanced orthogonal learning-driven multi-swarm sine cosine optimization: framework and case studies,” Expert Systems with Applications, vol. 144, Article ID 113113, 2020.
- A. F. Ba, H. Huang, M. Wang et al., “Levy-based antlion-inspired optimizers with orthogonal learning scheme,” Engineering with Computers, 2020.
- S. Jiao, G. Chong, C. Huang et al., “Orthogonally adapted Harris hawks optimization for parameter estimation of photovoltaic models,” Energy, vol. 203, Article ID 117804, 2020.
- Z. Xu, Z. Hu, A. A. Heidari et al., “Orthogonally-designed adapted grasshopper optimization: a comprehensive analysis,” Expert Systems with Applications, vol. 150, Article ID 113282, 2020.
- G. Xiong and D. Shi, “Orthogonal learning competitive swarm optimizer for economic dispatch problems,” Applied Soft Computing, vol. 66, pp. 134–148, 2018.
- X. Dan and B. Yuan, “Partially adaptive array design based on orthogonal projection algorithm,” in Proceedings of Third International Conference on Signal Processing (ICSP’96), Beijing, China, October 1996.
- T. Yokota and M. Gen, “Solving for nonlinear integer programming problem using genetic algorithm and its application,” in Proceedings of the IEEE International Conference on Systems, Man and Cybernetics, pp. 1602–1609, San Antonio, TX, USA, October 1994.
- A. Hu, D. M. York, and T. K. Woo, “Time-dependent density functional theory calculations of molecular static and dynamic polarizabilities, cauchy coeffificients and their anisotropies with atomic numerical basis functions,” Journal of Molecular Structure: THEOCHEM, vol. 591, no. 1–3, pp. 255–266, 2002.
- G.-G. Wang, S. Deb, A. H. Gandomi, and A. H. Alavi, “Opposition-based krill herd algorithm with Cauchy mutation and position clamping,” Neurocomputing, vol. 177, pp. 147–157, 2016.
- G. Salvador, A. Fernández, J. Luengo, and F. Herrera, “Advanced nonparametric tests for multiple comparisons in the design of experiments in computational intelligence and data mining,” Experimental Analysis of Power Information Sciences, vol. 180, no. 10, pp. 2044–2064, 2010.
- J. Alcalá-Fdez, L. Sánchez, S. García et al., “KEEL: a software tool to assess evolutionary algorithms for data mining problems,” Soft Computing, vol. 13, no. 3, pp. 307–318, 2008.
- S. Mirjalili and A. Lewis, “The whale optimization algorithm,” Advances in Engineering Software, vol. 95, pp. 51–67, 2016.
- S. Mirjalili, “SCA: a sine cosine algorithm for solving optimization problems,” Knowledge-Based Systems, vol. 96, pp. 120–133, 2016.
- Z. Yuhang, L. Qi, and W. Yang, “Materialized view selection algorithm—CSSA_VSP,” in Proceedings of the 2010 Second International Conference on Computational Intelligence and Natural Computing, Wuhan, China, September 2010.
- T.-C. Lu and J.-C. Juang, “An evolutionary space search algorithm (ESSA) for global numerical optimization, decision and control,” in Proceedings of the 48h ieee conference on decision and control (cdc) held jointly with 2009 28th chinese control conference, Shanghai, China, December 2009.
- W. Zhu, P. Xu, T. Bui, G. Wu, and Y. Yang, “Energy-efficient cell-association bias adjustment algorithm for ultra-dense networks,” Science China Information Sciences, vol. 61, no. 2, 2017.
- S. Gupta and K. Deep, “A hybrid self-adaptive sine cosine algorithm with opposition based learning,” Expert Systems with Applications, vol. 119, pp. 210–230, 2019.
- Y. Ling, Y. Zhou, and Q. Luo, “Lévy flight trajectory-based whale optimization algorithm for global optimization,” IEEE Access, vol. 5, pp. 6168–6186, 2017.
- J. J. Liang, A. K. Qin, P. N. Suganthan, and S. Baskar, “Comprehensive learning particle swarm optimizer for global optimization of multimodal functions,” IEEE Transactions on Evolutionary Computation, vol. 10, no. 3, pp. 281–295, 2006.
- C. García-Martínez, M. Lozano, F. Herrera, D. Molina, and A. M. Sánchez, “Global and local real-coded genetic algorithms based on parent-centric crossover operators,” European Journal of Operational Research, vol. 185, no. 3, pp. 1088–1113, 2008.
- C. A. Coello Coello, “Constraint-handling using an evolutionary multiobjective optimization technique,” Civil Engineering and Environmental Systems, vol. 17, no. 4, pp. 319–346, 2000.
- C. A. C. Coello, “Theoretical and numerical constraint-handling techniques used with evolutionary algorithms: a survey of the state of the art,” Computer Methods in Applied Mechanics and Engineering, vol. 191, no. 11-12, pp. 1245–1287, 2002.
- C. A. Coello Coello and E. M. Montes, “Constraint-handling in genetic algorithms through the use of dominance-based tournament selection,” Advanced Engineering Informatics, vol. 16, no. 3, pp. 193–203, 2002.
- K. Deb, “An effcient constraint handling method for genetic algorithms,” Computer Methods in Applied Mechanics and Engineering, vol. 186, no. 2–4, pp. 311–338, 2000.
- Q. He and L. Wang, “An effective co-evolutionary particle swarm optimization for constrained engineering design problems,” Engineering Applications of Artificial Intelligence, vol. 20, no. 1, pp. 89–99, 2007.
- K. M. Ragsdell and D. T. Phillips, “Optimal design of a class of welded structures using geometric programming,” Journal of Manufacturing Science and Engineering, vol. 98, no. 3, pp. 1021–1025, 1976.
- J. N. Siddall, Analytical Decision-Making in Engineering Design, Prentice-Hall, Englewood Cliffs, NJ, USA, 1972.
- M. Canayaz and A. Karci, “Cricket behaviour-based evolutionary computation technique in solving engineering optimization problems,” Applied Intelligence, vol. 44, no. 2, pp. 362–376, 2016.
- X. Zhang, T. Wang, J. Wang, G. Tang, and L. Zhao, “Pyramid channel-based feature attention network for image dehazing,” Computer Vision and Image Understanding, vol. 197-198, Article ID 103003, 2020.
- Y. Li, W.-G. Cui, H. Huang, Y.-Z. Guo, K. Li, and T. Tan, “Epileptic seizure detection in EEG signals using sparse multiscale radial basis function networks and the Fisher vector approach,” Knowledge-Based Systems, vol. 164, pp. 96–106, 2019.
- Y. Li, J. Liu, Z. Tang, and B. Lei, “Deep spatial-temporal feature fusion from adaptive dynamic functional connectivity for MCI identification,” IEEE Transactions on Medical Imaging, vol. 39, no. 9, pp. 2818–2830, 2020.
- X. Zhang, M. Fan, D. Wang, P. Zhou, and D. Tao, “Top-k feature selection framework using robust 0-1 integer programming,” IEEE Transactions on Neural Networks and Learning Systems, pp. 1–15, 2020.
- Y. Zhang, R. Liu, X. Wang, H. Chen, and C. Li, “Boosted binary Harris hawks optimizer and feature selection,” Engineering with Computers, 2020.
- Q. Li, H. Chen, H. Huang et al., “An enhanced grey wolf optimization based feature selection wrapped kernel extreme learning machine for medical diagnosis,” Computational and Mathematical Methods in Medicine, vol. 2017, Article ID 9512741, 15 pages, 2017.
- H. Huang, X. Feng, S. Zhou et al., “A new fruit fly optimization algorithm enhanced support vector machine for diagnosis of breast cancer based on high-level features,” BMC Bioinformatics, vol. 84, no. S8, 2019.
- X. Xue, S. Wang, L. Zhang, Z. Feng, and Y. Guo, “Social learning evolution (SLE): computational experiment-based modeling framework of social manufacturing,” IEEE Transactions on Industrial Informatics, vol. 15, no. 6, pp. 3343–3355, 2019.
- Y. Zhou, L. Tian, C. Zhu, X. Jin, and Y. Sun, “Video coding optimization for virtual reality 360-degree source,” IEEE Journal of Selected Topics in Signal Processing, vol. 14, no. 1, pp. 118–129, 2020.
Copyright © 2020 Hanfeng Zhou et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.