Abstract

A multi-offspring improved real-coded genetic algorithm (MOIRCGA) using the heuristical normal distribution and direction-based crossover (HNDDBX) is proposed to solve constrained optimization problems. Firstly, a HNDDBX operator is proposed. It guarantees the cross-generated offsprings are located near the better individuals in the population. In this way, the HNDDBX operator ensures that there is a great chance of generating better offsprings. Secondly, as iterations increase, the same individuals are likely to appear in the population. Therefore, it is possible that the two parents of participation crossover are the same. Under these circumstances, the crossover operation does not generate new individuals, and therefore does not work. To avoid this problem, the substitution operation is added after the crossover so that there is no duplication of the same individuals in the population. This improves the computational efficiency of MOIRCGA by leading it to quickly converge to the global optimal solution. Finally, aiming at the shortcoming of a single mutation operator which cannot simultaneously take into account local search and global search, a Combinational Mutation method is proposed with both local search and global search. The experimental results with sixteen examples show that the multi-offspring improved real-coded genetic algorithm (MOIRCGA) has fast convergence speed. As an example, the optimization model of the cantilevered beam structure is formulated, and the proposed MOIRCGA is compared to the RCGA in optimizing the parameters of the cantilevered beam structure. The optimization results show that the function value obtained with the proposed MOIRCGA is superior to that of RCGA.

1. Introduction

The genetic algorithm (GA) was proposed by Professor Holland and his students at the University of Michigan at the end of the 1960s and in the early 1970s [14]. In 1975, De Hong first proposed the evolutionary strategy (ES) of elitism preservation in his doctoral thesis. Later, others further studied it and proposed several elitism preservations and ES of replacing copying with selection [58]. At present, the GA is basically calculated according to these evolutionary strategies. In 1989, Goldberg [9] made a comprehensive and systematic summary and discussion of the GA and laid the foundation of the modern GA.

Initial GA uses binary coding. Because it can only map to discrete values in the search space and has Hamming distance, the accuracy of solution is not high [10]. In addition, binary coding needs to be encoded and decoded frequently, thus increasing the calculation time, with potential conversion error. Since the accuracy of the solution is controlled by the encoding length, for high accuracy, binary coding may need too long codes, resulting in excessive computing and memory space as well as reduced computational speed. In 1989, Lucasius et al. first proposed real-coded genetic algorithms (RCGAs) [11, 12]. RCGA has many attractive properties such as high precision, no coding and decoding required, effective large space search, simple computing, fast convergence, and not easy to fall into a local extreme value [13, 14].

GA has been widely applied in areas such as function optimization, automatic control, machine learning, combinatorial optimization, production scheduling problems, image processing, self-adaptation control, planning and design, industrial engineering, intelligent manufacturing systems, bioengineering, system engineering, artificial intelligence, intelligent machine system, artificial life, text information filtering, and cooperative multi-objective evaluation [1522]. It is especially suitable to deal with complex and nonlinear continuous optimization problems that cannot be solved by traditional search methods [6].

During the last few decades, the RCGA has attracted a lot of attention because of its unique and excellent performance. Many scholars have made in-depth research on RCGA and achieved excellent research results [2331]. Because crossover operators and mutation operators have a great influence on the performance of GA, many scholars have focused their attention on the improvement of crossover operators and mutation operators. In 1992, Wright [32] proposed a heuristic crossover (HX) operator, which is used to solve constrained and unconstrained optimization problems and achieved good results. The two offsprings generated by the HX operator are located on the straight line connecting the two parents, and are in the vicinity of the parent individuals with better fitness value. In 1993, Eshelman et al. [23] used the concept of interval schemata to develop a blend crossover operator (BLX-α). When the difference between the parents is small, the offsprings are similar to the parents, but if the difference between the parents is large, the generation of the offspring is similar to a random search. The parameter α is used to control the search domain, and the optimal value of α is suggested to be 0.5; BLX-α generates the offspring solution at the center from both selected parents. In 1995, Deb et al. [33] proposed a simulated binary crossover (SBX) which simulates the single-point binary crossover and generates two offsprings from two selected parents. The two offsprings generated by the SBX operator are located on the straight line connecting the two parents and are in the vicinity of the two parents. The distribution index η in the SBX operator controls the distance between two offsprings and two parents. If the value of η is large, the probability that the generated offsprings are closer to the two parents is greater; if the value of η is relatively small, the probability that the generated offsprings are farther away from the two parents is greater. The disadvantage of SBX is that it cannot adaptively control the size of the parameter η value, and therefore cannot adaptively control the distance between two offsprings generated by the crossover and the two parents. In 1997, Ono et al. [34] proposed the unimodal normal distribution crossover (UNDX) operator which used an ellipsoidal probability distribution to produce two or more offsprings from three selected parents. In this scheme, the offsprings located near the center of the first two parents are produced with a higher probability whereas those neighboring the parents are produced with a lower probability. In 2002, Deb et al. [35] developed a parent-centric crossover (PCX) operator. PCX is a self-adaptive multi-parent crossover which uses a large probability rather than the center of selected parents to generate a new solution near each parent. One parent is selected for each generated offspring and a difference vector is calculated between this parent and the N chosen parents. Besides, the PCX has difficulties to solve multimodal problems. The comparison results show that RCGA based on PCX outperforms other comparative crossover schemes in finding the global optimum for a number of test problems. In 2007, Deep et al. [36] proposed a Laplace crossover (LX) in which Laplace distribution is used as the density function to decide the location of the offspring. The two offsprings generated by the LX operator are symmetrical relative to their parental position, and the two offsprings are not necessarily located near the better among the two parents. In 2016, Chuang et al. [27] proposed a direction-based crossover (DBX) which is able to explore 2n − 1 possible search directions. As a result, a high probability to produce better offspring individuals is assured. However, the search directions of DBX are limited. Although it may generate a crossover direction capable of guiding the chromosomes to move towards the optimal solution; this is not highly probable. Meanwhile, when the dimensionalities of the variables are small, the null vector solution is likely to be generated. Currently, the search direction is randomly generated, and the search direction is not instructive. In 2018, Wang et al. [37] developed a heuristic normal distribution crossover (HNDX) operator. It can guarantee the cross-generated offspring to locate closer to the better one among the two parents and the search direction to be very close to the optimal search direction or to be consistent with the optimal searching direction. However, HNDX does not consider whether there are better individuals in the population than the two parents of participation crossover, and if so, whether it should be considered in the crossover operator.

Similarly, several other mutation schemes are reported in the literature. In 1996, Michalewicz [38] proposed a non-uniform mutation (NUM) operator. At the beginning of the iterations, NUM’s global search ability is strong; in the latter part of the iterations, NUM’s local search ability is strong. The disadvantage of the NUM operator is that it is difficult to choose a maximum number of iterations that applies to all problems. In 1995, Hinterding [39] proposed a Gaussian mutation operator in which the global and local search capabilities of GA are changed by adjusting the variance of the Gaussian mutation operator. The disadvantage of Gaussian mutation is that the local search ability is stronger than the global search ability, and the algorithm is easy to fall into local optimum. Therefore, Gaussian mutation is not suitable for solving multimodal optimization problems. In 1996, Wang et al. [40] proposed a mutation operator which mutated towards the gradient direction of the objective function. However, when the objective function is not differentiable, the mutation operator cannot be performed. In 2014, Deb et al. [41] proposed a polynomial mutation (PM) operator, which is widely used in RCGA. PM operators have strong random search ability and are not easy to fall into local optimum, but the convergence speed of the algorithm is slower. In 2016, Chuang et al. [27] proposed a dynamic random mutation (DRM) operator. Here, the calculation formula of the step size contradicts its interpretation. In addition, the mutation operator needed the maximum number of iterations in advance, which is difficult to estimate.

In the improved RCGA mentioned above, two parents generate two offsprings by crossover. In the process of biological evolution, a pair of parents usually breed more than two offsprings. Such species not only survive well in the process of evolution, but also make the species evolve so as to get better species. In fact, there is no species in which the number of offsprings bred by a pair of parents is less than or equal to two. Even if such species exist, in the process of evolution, they will eventually become extinct due to diseases, lack of food, and other factors. In addition, when the number of offsprings within a population is greater than that of the parents, the population size will expand. If the survival space of the species is constant, the competition within the population will aggravate. In this case, excellent individuals with strong viability within the population are easy to survive, and individuals with lower viability are eliminated, so that the species gradually evolves to adapt to the living environment and get better species. Therefore, during the evolution of species in nature, the number of offsprings should be greater than that of the parents.

Inspired by the laws of biological evolution in nature, in 2015, Wang et al. [42] proposed a multi-offspring GA with single-point crossover and verified its effectiveness of single-point crossover multi-offspring GA by using test functions. In 2016, Wang et al. [43] developed multi‐offspring GA. In this, the theoretical basis of multi-offspring GA was given. The schema theorem proved that multi-offspring GA is better than non-multi-offspring GA. This paper is only suitable for solving TSP. In 2016, Wang et al. [44] proposed a multi-offspring GA of two-point crossover and verified its validity using test functions. The multi-offspring GA given in the literature is mainly divided into two categories: one is the multi-offspring GA for solving the TSP; the other is the multi-offspring GA with binary coding.

In summary, the two offsprings generated by the LX operator are symmetric with respect to the parental position, and the two offsprings are not necessarily located near the better individuals of the two parents. The two offsprings generated by the SBX operator are located on the straight line connecting the two parents, and are in the vicinity of the two parent individuals. The two offsprings generated by the HX operator are located on the straight line connecting the two parents, and are in the vicinity of the parent individuals with better fitness value. The offspring individuals generated by PCX operators are not close to the center of the parental individuals, but close to the parental individuals. The two offsprings generated by the BLX-α operator locate at the center of the two parents of participation crossover. Since the DBX operator is only able to explore 2n − 1 possible search directions, the search directions of DBX are limited when the dimensionalities of the variables are small, and the null vector solution is likely to be generated. Currently, the search direction is randomly generated, and the search direction is not instructive. In order to overcome the shortcomings of the above crossover operator, this paper develops a heuristic normal distribution and direction-based crossover (HNDDBX) operator. HNDDBX not only considers the optimal individuals in the population, but also ensures that the offsprings generated after crossover are located near the better parent among the parents of participation crossover, not only near the better parent on the straight line connecting the two parents. In addition, a single mutation operator is difficult to take into account both global and local search abilities. For this reason, this paper develops a method of Combinational Mutation which includes three-mutation operators. Furthermore, because multi-offspring GA is superior to non-multi-offspring GA, the number of offsprings generated by the HNDDBX operator is more than those of the parents. Taking the above factors into consideration, this paper proposes a MOIRCGA, and it is compared with other GAs in the literature. The comparison results show that the performance of MOIRCGA is better than those of other GAs in the literature. Finally, MOIRCGA is applied to the optimization of the cantilever beam structure, and good results are obtained.

2. Penalty Function Method for Constrained Optimization Problems

The mathematic model of a constrained optimization problem can be generally expressed as follows:where n is the population size, hi(X) = 0 is the i-th equality constraint, p is the number of equality constraints, is the j-th inequality constraint, q is the number of inequality constraints, and Xk is a D-dimensional vector Xk = (xk1, xk2, …, xkD).

Equation (1) can also be expressed as

Stating is the optimal solution to the constrained optimization problems means that . In addition, if , the constraint is referred to as the active constraint. Under this concept, all the equation constraints hi(X) = 0 (i = 1, 2, …, p) are active at .

The penalty function method converts a constrained optimization problem to an unconstrained optimization problem by using two penalty factors and defining the penalty function to be minimized as [45]:where M1 and M2 are the penalty factors, generally chosen as positive constants that are big enough; the second and third terms on the right are the penalty terms, and P(X, M) is the penalty function.

In equation (3), when X ∈ R, there should be no penalty to the feasible points, thus P(X, M) = f(X); when X ∉ R, for the nonfeasible points, M1 and M2 should be very big. Moreover, when point X gets farther away from the feasible region, the penalty should be larger.

The minimum value of equation (3), that is,is equivalent to the minimum value of equation (1).

3. Multi-offspring Improved Real-Coded Genetic Algorithm (MOIRCGA)

In the multi-offspring GA, the number of offsprings is an integer multiple of the number of parents. Other GAs will be called conventional GAs.

Letting n0 and n1 be the number of parents and offsprings, respectively, we get

In the paper, the number of offsprings generated by crossover will be chosen as 2n. In comparing multi‐offspring GA with conventional GA, the competition among individuals in the population in multi‐offspring GA is more intensifier. Most of the surviving individuals in the population are excellent. This causes multi-offspring GA to have faster convergence speed than conventional GA.

3.1. Mathematical Symbols of MOIRCGA

We define the following symbols: is the initial population, where , i = 1, 2, …, n; D and n are the dimensions of the variables and the population size, respectively; a = [a1, a2, …, aD] and b = [b1, b2, …, bD] are lower and upper bounds of the variables . In the t-th iteration, Xi(t) is also referred to as a chromosome with known as genes. We also define Pc and Pm as the crossover mutation probabilities, respectively, and m is the number of preserved elitism individuals.

3.2. Initialization of Population

Before the population is initialized, the upper and lower bounds of the variables are determined. Hereafter, the variables are randomly generated uniformly within the interval [a, b]. Therefore, the i-th individual in the initial population is initialized aswhere (b − a).rand(1, D) is the product of multiplying (b − a) with the element at the same position of rand(1, D), and rand(1, D) represents D uniformly distributed random numbers in [0, 1].

3.3. ES of MOIRCGA

In the population evolution process of MOIRCGA, the population size is assumed to be constant. The ES of MOIRCGA is as follows:(1)Generate the initial population of size n, and sort all individuals in descending order according to their objective function values.(2)Select m elitist individuals with best ranking among n parents. Use the method of sorting grouping selection (SGS) to select the parents of participation crossover as detailed in Section 3.4.(3)Let the crossover probability Pc equal 1 and n parents generate 2n offsprings by crossover as detailed in Section 3.5.(4)Generate a new population consisting of 2n offsprings and m elitist individuals chosen in step 3.(5)Sort the individuals of the new population in descending order according to their objective function values.(6)Select m elitist individuals with the best ranking in the population.(7)Using the mutation operator, modify the 2n offsprings generated by crossover. This is discussed in Section 3.6.(8)Regenerate the new population consisting of the 2nPm mutated offsprings, 2n(1 − Pm) unmutated offsprings, and m elitist individuals chosen in step 6.(9)Sort the individuals of the regenerated new population in descending order according to their objective function values.(10)If the iterative termination condition is met, output the optimal solution and the optimal value. Otherwise, go to the next step.(11)Select n individuals with best ranking within the last population. Go to step 2 to start the next iteration.

The ES flow diagram for MOIRCGA is shown in Figure 1.

In the ES above, because the crossover probability is 1, the number of new individuals generated by crossover increases, and thus the possibility of generating excellent individuals increases. In addition, the number of new individuals generated by mutation is 2nPm, which is twice the number of new individuals generated by conventional GA mutation. Therefore, the possibility of obtaining excellent individuals and jumping out of local optimum is increased. This helps to improve the convergence speed of MOIRCGA.

3.4. Sorting Grouping Selection (SGS)

We assume the following: (1) the maximum of the objective function is sought, (2) the population size n is an even number, and (3) the individuals in the population are sorted with respect to their objective values P(X, M) in descending order. If the minimum of the objective function is sought, the objective function P(X, M) can be changed to the maximum value by maxP(X, M) = −min[−P(X, M)]. The population before sorting is X(t) = (X1(t), X2(t), …, (Xn(t)), and after sorting in descending order becomes X′(t) = (X1(t), X2(t), …, (Xn(t)), where P((X1(t), M) ≥ P((X2(t), M) ≥⋯≥ P((Xn(t), M), t being the number of iterations.

The individuals in the population are divided into two groups according to their objective function values. Group 1 includes the better n/2 individuals, that is, X1(t), X2(t), …, Xn/2(t), and Group 2 includes the other n/2 individuals, that is, Xn/2+1(t), Xn/2+2(t), …, Xn(t). The 1st individual in Group 1 is matched with the 1st individual in Group 2, the 2nd individual in Group 1 is matched with the 2nd individual in Group 2, and so on. In this way, we obtain 0.5n pairs of individuals of participation crossover. This selection method is called SGS. The advantage of the SGS method is that the selection operation is completed according to the objective function value of each individual, and it is not necessary to calculate the fitness value of each individual. This causes selection operation speed to be fast.

3.5. Heuristic Normal Distribution and Direction-Based Crossover (HNDDBX)

If the optimization problem is to solve for the maximum value of the objective function, the optimal solution is likely to be near the individual with a bigger objective function value [46]. Hence, the better individuals in the population are more likely to approach the optimal solution. Inspired by this, this paper develops a heuristic normal distribution and direction-based crossover (HNDDBX) operator.

The key to HNDDBX is how to make the crossover generated offspring be in the vicinity of the better individuals in the population. Since a random number generated by the normal distribution has a higher probability of being located near its mean μ, if a better individual in the population is used as being the mean μ of the normal distribution, then it is guaranteed that its offspring according to normal distribution is in the vicinity of the better individual. Therefore, a normal distribution can be used to generate offsprings. The normal distribution is denoted by N(μ, σ2), where μ is the mean, and σ2 is the variance. The density function of the normal distribution is shown in Figure 2.

It can be seen from Figure 2 that the probability of a random number X generated by N(μ, σ2) in the interval (μ − σ, μ + σ) is 68.26%, the probability of X in the interval (μ − 2σ, μ + 2σ) is 95.44%, and the probability of X in the interval (μ − 3σ, μ + 3σ) is 99.72%. It can be seen that the probability that X falls outside the interval (μ − 3σ, μ + 3σ) is less than 0.3%. Thus, the interval (μ − 3σ, μ + 3σ) can be regarded as the actual possible interval of the random variable X, which is called the 3σ principle of the normal distribution. Based on the above analysis, as long as we control the size of σ value, we can basically guarantee that the random number X generated by N(μ, σ2) is near the mean μ. If the mean μ of the normal distribution is equal to the better individual X within the parents of participation crossover, it can be ensured that the individuals generated by the normal distribution are located near the better individual X.

In order to explain the HNDDBX operator, we assume that the population size n is even. After sorting in descending order according to objective function values, the population is X′ = (X1, X2, …, Xn). At the t-th iteration, the optimal individual in the population is X1. This paper uses the SGS method to select individuals of participation crossover. Let the paired parents of participation crossover be Xi (i = 1, 2, …, n/2) and Xj (j = n/2 + 1, n/2 + 2, …, n). Assuming that Xi is superior to Xj, the offsprings are generated bywhere D is the dimension of the variables, Yi (i = 1, 2, …, 2n) are the offsprings generated by the crossovers, r11, r12, …, r1D and r21, r22, …, r2D are uniformly distributed random numbers in [0, 1], ε is a small positive number to avoid that the variance of the normal distribution is 0. means that the elements at the same position in two matrices or vectors are multiplied.

To illustrate the principle of HNDDBX, let the variable dimension D be 2. At the t-th iteration, let the two parent individuals of participation crossover be Xi and Xj, respectively. If Xi is superior to Xj, X1 is the best individual in the population, and is the optimal solution to the problem to be solved. The principle of HNDDBX is shown in Figure 3.

In Figure 3, and . Equation (7) shows that is the center of the better individuals in the population, is the best individual in the population, is the better one among the two parent individuals of participation crossover, and is the center of , , and . So, in most cases, is superior to . Offsprings Yi (i = 1, 2, …, n/2) are randomly generated according to , with its mean equal to the mean of and its variance equal to . In addition, it can be seen from Figure 3 that the random number generated according to has high probability of being in the vicinity of the mean of the normal distribution. According to the above analysis, the two parent individuals and generate offsprings Yi, which have a great possibility to be located within a circle with as its center and as its radius, as shown in Figure 3. When the dimension of the variables is greater than 2, Yi has a great possibility to be located within a sphere with as its center and as its radius. Based on the above analysis, as long as we control the size of variance , the HNDDBX operator can basically guarantee that the offsprings generated by crossover are near the mean . Similarly, the offspring individual Yj (j = n/2 + 1, n/2 + 2, …, n) generated by crossover according to equation (7) is near the optimal individual in the population. Because is the best individual in the population, the offsprings generated by crossover with the HNDDBX operator are expected to be better than those of the crossover operator in [31, 37].

It can be seen from Figure 3 that is the search direction for Yk (k = n + 1, n + 2, …, 1.5n). When r12 = 0, since r11 is a random number uniformly distributed in [0, 1], and the search direction is , Yk is uniformly distributed on the line segment . When r11 = 0, because r12 is a random number uniformly distributed in [0, 1] and the search direction is , Yk is uniformly distributed on the line segment . When r11 = r12, since r11 and r12 are random numbers uniformly distributed in [0, 1] and the search direction is , Yk is uniformly distributed on the line segment . When r11 and r12 are random numbers uniformly distributed in [0, 1], the search directions are any directions between and . At this point, there are countless search directions, and Yk may be located at any point within . In addition, because is better than , the offsprings Yk generated by the HNDDBX operator have a great possibility to be superior to those of the crossover operator in [31, 37]. Thus, Yk may be very close to the optimal solution of the problem to be solved. When r22 = 0, since r21 is a random number uniformly distributed in [0, 1] and the search direction is , Yl is uniformly distributed on the line segment . When r21 = 0, because r21 is a random number uniformly distributed in [0, 1] and the search direction is , Yl is uniformly distributed on the line segment . When r21 and r22 are random numbers uniformly distributed in [0, 1], the search directions are any directions between and . At this point, there are countless search directions, and Yl may be located at any point within ; Yl may be very close to the optimal solution of the problem to be solved.

In summary, the offsprings generated by the HNDDBX crossover operator are better than those of the crossover operator in [31, 37]. Therefore, the HNDDBX operator can significantly improve the convergence speed of MOIRCGA.

As an example, in terms of 2D search space, for the spatial positions of two parent individuals Xi and Xj participating in crossover, there are three possible cases: (a) both parents are in the infeasible region; (b) one parent is in the feasible region while the other is not; and (c) both parents are in the feasible region. Of these three cases, the HNDDBX operation in a two-dimensional space is shown in Figure 4.

In Figure 4, the schematic of case (a) shows that the HNDDBX operator can force two parents to move from the infeasible region to the feasible region. In the case of (b), the HNDDBX operator enables the individuals in the infeasible region to move into the feasible region, and the individuals that have been in the feasible region improve their objective function values by HNDDBX. When the individuals are already in the feasible region, as in the case of (c), HNDDBX searches near the better individual and the optimal individual moves towards the global optimal solution.

3.6. Substitution Operation

In the global optimization problem with many local optima, when the MOIRCGA finds a region with an extreme value (whether it is a local extremum or a global extremum), individuals in the population constantly move closer to the region, and they may be the same or similar individuals in the population. With the increase of the iterations, the same individuals in the population will gradually increase and may even make all the individuals in the population to be the same. If there are many such individuals in the population, it is likely that the two parents of participation crossover are the same, such that the offsprings Yk generated by crossover, according to equation (7) are the same as the two parents of participation crossover. Hence, the crossover operation does not work, and GA is unable to converge to the global optimal solution. In order to avoid the above phenomenon, a substitution operation is added after crossover operation of GA. The method of substitution operation is as follows: the population size after crossover is 2n; if there are two or more same individuals in the population, only one of those is retained, and the remaining individuals are removed. Let the population size be n1 at this time. In order to maintain the population size after the crossover as 2n, 2n − n1 individuals are randomly generated, and 2n − n1 same individuals are replaced by randomly generated 2n − n1 individuals. The following examples illustrate the substitution operation.

Suppose the population size is 10. After several iterations, the population is X. At this time, there are same individuals in the population. In order to improve the efficiency of crossover operation and avoid falling into local optimum, only one of the same individuals is retained; the population after removing the same individuals is X1, which has four fewer individuals than X. In order to keep the size of the population unchanged, 4 individuals Y are randomly generated. The population X′ after the substitution operation consists of X1 and Y. X, X1, and X′ are shown in Table 1.

3.7. Combinational Mutation

With the mutation operators given in the existing literature, some local search abilities are strong [47, 48], such as Gauss mutation operator, and some global search abilities are strong [48], such as Cauchy mutation operator. For optimization problems with less number of extreme points, a mutation operator with stronger local search ability should be adopted. For optimization problems with more number of extreme points, if a mutation operator with stronger local search ability is adopted, it is easy to converge to the local optimum; if a mutation operator with strong global search ability is adopted, and the accuracy requirement of the optimal solution for the problem to be optimized is higher, the convergence speed of the algorithm slows down. With some of the mutation operators given in the literature, there is strong global search capability at the beginning of the iterations, and with the increase of the number of iterations, local search capabilities are enhanced. However, this requires the maximum number of iterations to be given, which is difficult to do in advance. Although this kind of mutation operator is theoretically feasible, it actually has poor performance [27]. In summary, a single mutation operator is difficult to take into account both global and local search abilities. For this reason, we developed a method of combinational mutation which includes three-mutation operators. They are as follows.

The first mutation operator is a mutation operator based on the Cauchy distribution called the Cauchy mutation operator. The Cauchy mutation operator is as follows:where Cauchy (0, 1) is the standard Cauchy distribution, is the individual to be mutated, and is the individual after the mutation.

The second one is the mutation operator based on the normal distribution called the normal mutation. The normal mutation operator is as follows:where μ is the mean of the normal distribution, usually chosen as μ = , and δ2 is the variance of the normal distribution.

δ2 of (9) is given bywhere Xbest is the optimal individual in the population.

The third one is the mutation operator based on Lévy flight, called the Lévy mutation. It is as follows:where α is the step size scaling factor, generally chosen as α = 0.01.

The calculation formula for the simulated Lévy flight path proposed by Mantegna is as follows [49]:where 0 < λ < 2, usually chosen as λ = 1.5.where is the Gamma function.

The two-dimensional space Lévy flight diagram is shown in Figure 5.

As seen in Figure 5, the characteristics of Lévy flight are as follows: (1) there are a lot of small steps, that is, most of the time, it gives a local search; (2) sometimes there is a large displacement, so that individuals in the population do not search only in one place, resulting in occasional global search capability.

The specific method of combinational mutation is as follows: in the t-th iterations, we divide the number of iterations by 3. When the remainder is 1, the first mutation operator is used; when the remainder is 2, the second mutation operator is used; when the remainder is 0, the third mutation operator is used.

Compared with the normal mutation operator, the generated offspring of Cauchy mutation operator has high probability of being far away from the individual to be mutated, so the global search ability of Cauchy mutation operator is strong. The normal mutation operator focuses on searching for a local region near the individual to be mutated, and the local search ability is stronger, but the ability to guide the individual to jump out of the local extremum is weak, which is not conducive to global convergence. The Lévy mutation operator performs local searches in most cases but occasionally performs global searches. Therefore, the advantage of combinational mutation is that both the local search ability and the global searching ability are taken into account.

3.8. Pseudocode of MOIRCGA

The pseudocode of MOIRCGA is shown in Algorithm 1.

Input: population size: n; mutation probability: Pm; crossover probability: Pc; lower and upper bounds of variables: [a, b]; the number of elite individuals retained: m
Output: the optimal value of an optimization problem
(1)Initialize population
(2)Calculate the objective function values of individuals and sort the individuals in descending order
(3)While termination criterion is not met
(4)Perform SGS
(5)Perform HNDDBX
(6)Perform substitution operation
(7)Calculate the objective function values of individuals and sort the individuals in descending order
(8)Retain m elitist members from all parents and 3n offsprings after performing the substitution operation
(9)Perform combinational mutation
(10)Calculate the objective function values of the individuals and sort the individuals in descending order
(11)m elitist members, 3nPm mutated members, and 3n(1 − Pm) nonmutated members constitute the current population
(12)Select n best members to constitute the new population, and retain m elitist members
(13)End While
(14)Output the average running time, the average number of iterations, and the optimal value

4. Algorithmic Testing and Analysis

RCGA proposed in [27, 31, 37] is abbreviated as IRCGA-1, IRCGA-2, and IRCGA-3, respectively. The single-point crossover multi-offspring GA proposed in [42] is abbreviated as SPXMOGA, and hybrid GA in [50] is abbreviated as HGA.

In order to verify the validity and feasibility of MOIRCGA, it is compared with IRCGA-1, IRCGA-2, IRCGA-3, SPXMOGA, and HGA.

4.1. Iteration Termination Condition

The iteration termination condition for GA is to be defined aswhere is the true global optimal value of the i-th test function, fi is the optimal value of the i-th test function obtained by GA, and εi is the error for the i-th test function.

4.2. Test Functions

Sixteen well-known benchmark constrained functions were selected and are described in Appendix A [27, 31, 37].

4.3. Parameter Settings

In order to obtain a fair performance comparison, the parameters are set as follows.

(1) The computing precision of various test functions: ε1 = ε2 = ε3 = ε4 = ε5 = ε6 = ε7 = ε8 = ε9 = ε11 = ε12 = ε13 = ε14 = ε15 =ε16 = 10−4, ε10 = 10−2; (2) the population size n = 100; (3) M1 = 109 and M2 = 107 in the penalty function; (4) in MOIRCGA, the mutation probability Pm = 0.5 and the crossover probability Pc = 1; (5) in IRCGA-1, IRCGA-2, IRCGA-3, SPXMOGA, and HGA, the parameter settings are the same as in [27, 31, 37, 42, 50]; (6) reserve 50 elite individuals in MOIRCGA; and (7) the ES of IRCGA-1, IRCGA-2, IRCGA-3, and SPXMOGA are the same as that of in [27].

4.4. Comparison with the Results of Other Studies in the Literature

16 test functions in Appendix A were used as examples and each test function was run on the same computer for 1000 times. The initial population of the various algorithms is the same.

All of the algorithms in this paper were developed in Matlab R2018b programming language. IRCGA-1 [27] proposed a direction-based heuristic crossover operator which is also used in this paper. IRCGA-2 [31] gives a selection method and a direction-based heuristic crossover operator. IRCGA-3 [37] gives a heuristic normal distribution crossover (HNDX) operator. However, the HNDX does not consider the influence of the optimal individual in the population on the offsprings. SPXMOGA [42] proposes a single-point crossover multi-offspring GA based on binary coding and gives the generation method of multi-offspring and the corresponding evolutionary strategy, where the crossover operator is not heuristic. A new hybrid GSA-GA algorithm is presented in [50], and a HNDX operator is used in program. All of the above algorithms have parallelism, and regardless of whether the problem to be optimized has a derivative, it can be solved by the above algorithms. The computational results of various algorithms are shown in Table 2.

The average running time and the average number of iterations at convergence to the optimal solution were calculated as follows: when the iteration termination condition is satisfied, the numbers of iterations and time of processing at the i-th run are, respectively, iter(i) and t(i), i = 1, 2, …, k, where k is the number of runs used in each experiment. Then, the average running time and the average number of iterations are computed bywhere tave is the average running time and iterave is the average number of iterations.

When the average running time of ICGA-1 is t1, the average running time of MOIRCGA is t2; MOIRCGA reduces the average running time by x% for the i-th test function fi in comparison to IRCGA-1. x% is computed by

The computational method of x% is also the same as equations (16) and (17) for IRCGA-2, IRCGA-3, SPXMOGA, and HGA.

It is observed in Table 2 that the average running time and the average number of iterations of MOIRCGA are significantly superior to those of IRCGA-1, IRCGA-2, IRCGA-3, SPXMOGA, and HGA. MOIRCGA reduces the average running time by 86.4723%, 78.1382%, 77.1723%, 92.3419%, and 7.7381% for f1, 92.9825%, 81.4815%, 68.7500%, 94.8586%, and 51.2195% for f2, 92.5178%, 33.6842%, 47.5000%, 92.3544%, and 7.3529% for f3, 92.9012%, 36.1111%, 47.7273%, 96.2388%, and 65.9259% for f4, 76.2048%, 15.0538%, 16.8421%, 88.0030%, and 11.7318% for f5, 73.8462%, 56.4103%, 51.7730%, 74.9077%, and 88.6855% for f6, 84.8276%, 85.3982%, 84.6154%, 89.7727%, and 77.0302% for f7, 60.3208%, 55.6854%, 45.9896%, 81.7335%, and 41.9980% for f8, 93.1741%, 78.0220%, 66.3866%, 93.9668%, and 54.7170% for f9, 20.0993%, 56.8114%, 51.5437%, 62.8427%, and 15.8803% for f10, 70.5069%, 58.9744%, 53.6232%, 96.5720%, and 53.6232% for f11, 72.1264%, 70.2910%, 44.5714%, 91.5725%, and 31.4488% for f12, 91.5225%, 37.7119%, 12.5000%, 92.3676%, and 5.7692% for f13, 81.4116%, 15.8228%, 7.3171%, 87.5468%, and 7.9585% for f14, 79.3843%, 85.8515%, 78.6885%, 98.8135%, and 64.6965% for f15, and 48.7271%, 51.5362%, 46.5960%, 86.8151%, and 16.0934% for f16 in comparison to IRCGA-1, IRCGA-2, IRCGA-3, SPXMOGA, and HGA. Similarly, the number of iterations are reduced by 93.0655%, 90.8858%, 87.9022%, 99.0044%, and 63.7673% for f1, 88.8700%, 82.9963%, 64.0432%, 7.1735%, and 45.0541% for f2, 91.3308%, 76.9546%, 80.0369%, 84.7351%, and 35.6557% for f3, 84.5708%, 67.1235%, 72.3050%, 98.7247%, and 58.4379% for f4, 21.6585%, 42.5447%, 45.4985%, 95.1432%, and 11.8526% for f5, 68.5968%, 75.5886%, 73.4671%, 94.0034%, and 81.0207% for f6, 45.0158%, 90.6654%, 89.8940%, 91.3991%, and 38.1776% for f7, 31.3400%, 51.7993%, 71.8694%, 92.8838%, and 18.6654% for f8, 78.6502%, 84.2126%, 79.0909%, 86.8434%, and 23.1099% for f9, 28.9076%, 72.8941%, 70.3852%, 74.0789%, and 23.3931% for f10, 73.1964%, 78.7951%, 75.3992%, 83.8478%, and 18.9170% for f11, 61.3726%, 77.4691%, 65.1300%, 84.6915%, and 9.6795% for f12, 86.5586%, 69.7727%, 28.6090%, 89.1997%, and 18.1336% for f13, 33.0088%, 52.0363%, 37.0399%, 68.3551%, and 5.7958% for f14, 26.7832%, 88.9896%, 86.2090%, 89.4509%, and 12.6428% for f15, and 19.7828%, 57.7981%, 52.5073%, 80.9492%, and 32.3387% for f16. Thus, MOIRCGA is observed to be superior to IRCGA-1, IRCGA-2, IRCGA-3, SPXMOGA, and HGA in terms of all the measures utilized.

5. Parameter Optimization of Cantilever Beam

To further verify the validity of the MOIRCGA, the cantilever beam design problem with discrete cross-sectional area given in [50] was chosen as an application. The cantilever beam structure is shown in Figure 6.

The goal of the cantilever beam design optimization problem is to determine the optimal combination of five different cross-sectional areas to minimize the volume of the cantilever beam. The design problem has 10 variables. They are the width and height of each cross section, hi and bi (i = 1, 2, …, 5). An external force p = 50,000 N is applied at the free end of the cantilevered beam. The maximum allowable stress at the end of each section is σmax = 14,000 N/cm2, the material elasticity modulus E is 200 Gap, the length of each section li (i = 1, 2, …, 5) is 100 cm, and the maximum allowable deflection is ymax = 2.715 cm. The height-to-width aspect ratio of each cross-section is restricted to be less than 20. Then, the mathematical model for the optimization of this problem is defined as follows:

There are 11 constraints in this problem. Among them, are related to the allowable stress constraints, is the constraint regarding the allowable deflection, and are restrictions to the geometric shape of the cross-section. The optimization design problem in this paper is addressed by using the penalty function method given by equation (3). The parameter settings are as described in Section 4.3.

The optimization results of MOIRCGA are compared with the optimization results in reference [27, 31, 37, 42, 50, 51]. The optimization results (the best objective function values and design variable values) that are obtained with MOIRCGA and genetic algorithm (RCGA) in reference [27, 31, 37, 42, 50, 51] are listed in Table 3, and the constraints are listed in Table 4.

It is clear from Table 3 that the function value obtained with the proposed MOIRCGA is superior to that with RCGA in reference [27, 31, 37, 42, 50, 51]. At the same time, Table 4 reveals that the optimal solution obtained by the two methods satisfies all constraints.

6. Conclusions

The better individuals in the population are more likely to approach the global optimal solution. However, the two offsprings generated by the LX operator are symmetric with respect to the parental position, and the two offsprings are not necessarily located near the better individual of the two parents. The two offsprings generated by the SBX operator are located on the straight line connecting the two parents, and are in the vicinity of the two parents. The two offsprings generated by the Modified SBX-crossover (MSBX) operator are located on the straight line connecting the two parents, and are in the vicinity of the two parents. The distance between two offsprings and two parents can be adaptively controlled by parameter η, which overcomes the deficiency of adjusting distribution index η which SBX cannot adaptively adjust. The two offsprings generated by the HX operator are located on the straight line connecting the two parents and are in the vicinity of the parents with better fitness value. The offsprings generated by the PCX operators are not close to the center of the parents but close to the parents. The two offsprings generated by the BLX-α operator are located at the center of the two parents. Since the DBX operator is only able to explore 2n − 1 possible search directions, the search directions of DBX are limited; when the dimensionalities of the variables are small, the null vector solution is likely to be generated. Currently, the search direction is randomly generated, and the search direction is not instructive. HNDDBX overcomes the shortcomings of the above crossover operators. It not only considers the optimal individuals in the population, but also ensures that the offsprings generated by the HNDDBX operator are located near the better individual among the parents of participation crossover. Therefore, the convergence speed of the MOIRCGA using the HNDDBX operator is significantly improved.

In most previous GA algorithms, as iterations increase, the same individuals are likely to appear in the population. Therefore, it is possible that the two parents of participation crossover are the same. Under these circumstances, the crossover operation does not generate new individuals, that is, the crossover operation does not work, which affects the computational efficiency of GA and the ability to explore other extremal regions. To avoid this problem, the substitution operation is added after the crossover so that there is no duplication of the same individuals in the population. This improves the computational efficiency of MOIRCGA for quick convergence to the global optimal solution.

The local and global search abilities of different mutation operators are different. Some mutation operators have strong local search ability, and some have strong global search ability. A single mutation operator is difficult to take into account both local and global search abilities. This paper proposes a combinational mutation method which include three mutation operators, that is, Normal mutation operator, Cauchy mutation operator, and Lévy mutation operator. The Normal mutation operator has strong local search ability, and the Cauchy mutation operator has strong global search ability. The Lévy mutation operator performs local search in most cases and occasionally performs global search as well, but the global search ability of the Lévy mutation operator is superior to the Cauchy mutation operator, thus helping the algorithm avoid falling into local optimum.

The computational results with sixteen examples show that MOIRCGA has better performance than the other methods in references [27, 31, 37, 42] with respect to all the measures of performance considered.

As an example application, the optimization problem of the cantilevered beam structure is formulated, and MOIRCGA in the paper and RCGA in reference [52] are used to optimize the parameters of the cantilevered beam structure. The optimization results show that the function value obtained with MOIRCGA is superior to that obtained with RCGA.

Appendix

A. Test Functions

The 16 test functions are as follows:(1)f1:

The global optimal solution of f1 is located at with .

(2)f2:

The global optimal solution of f2 is located at with .

(3)f3:

The global optimal solution of f3 is located at with .

(4)f4:

The global optimal solution of f4 is located at and (0.0898, −0.7126) with .

(5)f5:

The global optimal solution of f5 is located at with .

(6)f6:

The global optimal solution of f6 is located at with .

(7)f7:

The global optimal solution of f7 is located at with .

(8)f8:

The global optimal solution of f8 is located at with .

(9)f9:

The global optimal solution of f9 is located at with .

(10)f10:

The global optimal solution of f10 is located at with .

(11)f11:

The global optimal solution of f11 is located at with .

(12)f12:

The global optimal solution of f12 is located at with .

(13)f13:

The global optimal solution of f13 is located at with .

(14)f14:

The global optimal solution of f14 is located at with .

(15)f15:

The global optimal solution of f15 is located at with .

(16)f16:

The global optimal solution of f16 is located at with .

Data Availability

The data used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest

The authors declare no conflicts of interest.

Acknowledgments

This work was supported by the project of National Key R&D plan (grant no. 2018YFD0300105).