Abstract

Cooperative coevolution (CC) is an effective framework for solving large-scale global optimization (LSGO) problems. However, CC with static decomposition method is ineffective for fully nonseparable problems, and CC with dynamic decomposition method to decompose problems is computationally costly. Therefore, a two-stage decomposition (TSD) method is proposed in this paper to decompose LSGO problems using as few computational resources as possible. In the first stage, to decompose problems using low computational resources, a hybrid-pool differential grouping (HPDG) method is proposed, which contains a hybrid-pool-based detection structure (HPDS) and a unit vector-based perturbation (UVP) strategy. In the second stage, to decompose the fully nonseparable problems, a known information-based dynamic decomposition (KIDD) method is proposed. Analytical methods are used to demonstrate that HPDG has lower decomposition complexity compared to state-of-the-art static decomposition methods. Experiments show that CC with TSD is a competitive algorithm for solving LSGO problems.

1. Introduction

Large-scale global optimization (LSGO) problems are widespread optimization problems in scientific research and engineering applications [16]. In this paper, continuous LSGO problems with 1000 or greater decision variables are studied and solved.

There are two main challenges in solving LSGO problems: (1) dimensional catastrophe [7] and (2) traditional mathematical algorithms and evolutionary algorithms (EAs) with rapid performance degradation as the dimensionality grows [810]. There are two main approaches to solve difficulties. The first approach is to design or improve EAs with powerful exploration so that the EAs avoid getting trapped in local optima in large-scale solution space [11]. The second approach is the CC with problem decomposition, which adopts the idea of divide-and-conquer and is divided into two stages [12, 13], problem decomposition and problem optimization. Recently, CC has been used to solve a variety of problems [1416]. The second approach is the one studied in this paper.

Decomposition methods decompose LSGO problems into multiple subproblems. CC with dynamic decomposition method decomposes the problems dynamically during the optimization stage, although dynamic decomposition methods are costly to decompose problems, which decompose fully nonseparable problems into multiple subproblems according to certain criteria [17]. The most typical dynamic decomposition method is random grouping [18], but the random grouping method is unable to make near-optimal decomposition for many variables. Other dynamic decomposition methods such as delta grouping lose effectiveness for partially separable problems with more than one nonseparable subcomponent [19], and CCEA-AVP is computationally expensive and effective only when detecting linear variable correlations without nonlinear interdependency [20].

CC with static decomposition method decomposes the problem in the problem decomposition stage, and advanced static decomposition method can accurately decompose problems [2123]. The key for static decomposition method is to correctly identify the relationships between variables. The differential grouping (DG) method [24] is an effective decomposition method for identifying relationships between variables; however, DG has the drawbacks of being sensitive to threshold and not being able to identify overlapping problems. The variant DG2 of DG effectively alleviates the drawbacks of DG, but DG2 is costly to compute adaptive threshold [25]. Mei et al. proposed a global differential grouping (GDG) to alleviate the sensitivity of DG to threshold by setting a global threshold [26]. Sun et al. proposed a recursive differential grouping (RDG) [27] and RDG’s variants RDG2 [28] and RDG3 [29] to further improve the performance of the DG decomposition problems. The recursive interaction structure in RDG, which significantly improves the variable interaction efficiency and the complexity of the variable interaction structure, is , where n denotes the problem dimension. Recently, Yang et al. proposed an efficient recursive differential grouping (ERDG) method to avoid repeated interactions during variable interactions [30]. However, the number of function evaluations (FEs) used to decompose problems increases quickly when the problem dimension rises; and the CC with MRF-based decomposition mechanism [31] proposed by Sun et al. and the three-level recursive differential grouping method [32] proposed by Bin et al. did not improve the ability of CC to solve nonseparable problems.

However, the static decomposition method aims to group the decision variables correctly, so CC with static decomposition encounters a dilemma in solving the fully nonseparable problems. To further reduce the decomposition complexity and alleviate the dilemma, the main work of this paper is as follows:(1)We propose a decomposition method TSD in this paper.(2)An improved differential grouping method, HPDG, is proposed. HPDG is a static decomposition method with a very low complexity. A new decomposition framework, hybrid-pool-based detection structure (HPDS), is proposed in this paper. The complexity of HPDS is , where n and m denote the problem dimension and the number of subproblems, respectively.(3)A unit vector-based perturbation (UVP) strategy is proposed to improve the decomposition accuracy of HPDG.(4)A known information-based dynamic decomposition (KIDD) method is proposed to decompose the fully nonseparable problems for alleviating the dilemma of CC with static decomposition in solving LSGO problems.

In the remainder of this article, we first introduce the differential grouping, variable interaction structure, and CC framework in Section 2. Then we present the CC with TSD in detail in Section 3. Numerical experiments are carried out and comparisons with the state-of-the-art algorithms are made in Section 4. Section 5 concludes the work of this paper and plans the next work.

2.1. Cooperative Coevolution with Decomposition Method

Cooperative coevolution plays an important role in the development of populations, especially for large size populations. In evolutionary computation, the effectiveness of CC is very dependent on the efficient dividing of population, which in the LSGO problems is problem decomposition [24, 3335].

The steps for CC to solve LSGO problems are described in Algorithm 1 [36]. holds the indexes of the decision variables of the subgroup. In traditional CC framework, computational resources are equally distributed, and line 5 is not executed. First, the population is initialized and static decomposition method decomposes LSGO problems (lines 1-2). Then each subgroup is optimized separately and the population is updated. To improve the utilization of computational resources, contribution-based CC (CBCC) is proposed as described in line 5 [34, 37]. After optimizing each subgroup once separately, the subgroup with the largest contribution value is selected to optimize to the termination condition. To balance the exploitation and exploration, the subgroup with the largest contribution is selected by CBCC3 [38], which is evolved to the condition where the subgroup with the second largest contribution from the last optimization is larger than the subgroup with the largest current contribution.

(1)initialize (population);
(2)subgroups group (FunID, D);
(3)fordo
(4) Optimize each subgroup separately;
(5) Select a subgroup with heaviest contribution and optimize to termination;
(6)end for
(7)return;
2.2. Decomposition Method
2.2.1. Problem Decomposition Framework

Problem decomposition framework is the logical structure of problem decomposition adopted by decomposition methods, and the logical structure determines the computational complexity of decomposition methods. Problem decomposition framework can be categorized into two types: double-pool decomposition framework and single-pool decomposition framework [39]. The double-pool decomposition framework includes a detection pool and an update pool, the identification between variables is performed in detection pool, and the identified variables are stored in the update pool, as shown in Figure 1(a). The variable interaction structure of the decomposition methods based on differential grouping belongs to the double-pool decomposition framework [24, 2729]. In the single-pool decomposition framework, the two pools are integrated into a single one, as shown in Figure 1(b), and the interactions of decision variables can be identified one by one. Experiments [39] show that single-pool decomposition framework is more conducive to identification of the variables than double-pool decomposition framework. In this paper, we combine the advantages of these two decomposition frameworks and propose a hybrid-pool decomposition framework.

2.2.2. Differential Grouping and Its Improvement

For differential functions, there are three relationships between decision variables: no relationship, direct relationship, and indirect relationship [24]. Based on these three relationships, three types of LSGO problems are constituted: full separable problem, partially separable problem, and full nonseparable problem, and the overlapping problem is a special type of partially separable problem. The primary task of problem decomposition is to identify the relationships between decision variables. For LSGO problem , the decision vector X of is divided into mutually exclusive subsets, as shown in equation (1), with for any .

DG decomposes problems by identifying the relationships between each pair of variables. Let the two variables identified be and , if the partial derivative is

In the above equation, is a very small positive real number parameter of DG used for determining whether two variables are related. Then is unrelated to ; otherwise is related to . If the relationship between two variables is indirect, it is difficult for the DG to identify this relationship, as in the problem described by the following equation:where and are related through interacting with each other and DG cannot find the relationship between and because

Problem decomposition by identifying the relationship between each pair of variables is inefficient. RDG significantly reduces the FEs by using a recursive detection structure in unidentified variables to complete problem decomposition. The theory of the identification for two sets of variables proved by Sun et al. is as follows [27]:

Let be subsets of the decision vector X of , let and be unit vectors of and , respectively, and is a feasible solution, where , , and . If some variables of and are related, inequality (7) holds. Conversely, and are not related to each other [27].

Based on RDG, ERDG uses interaction history information to complete problem decomposition and further reduces the FEs [27, 30]. However, RDG and ERDG apply perturbations to the variables by giving the same upper and lower bounds to the variables to be identified, so that the perturbation to identify all problems is fixed. The two changes of the fitness value arewhere and are candidate solutions that are given upper bounds and lower bounds, respectively, and mb denotes the mean of the upper bounds ub and lower bounds lb. The variables belonging to and are assigned the same upper bounds and lower bounds as , respectively, and the variables that do not belong to are assigned mb.

3. Proposed Method

3.1. Cooperative Coevolution with Two-Stage Decomposition

In this section, CC with TSD is proposed, which performs problem decomposition in the decomposition and optimization stages. In the decomposition stage, first the set of candidate solutions “population” is first initialized, and then HPDG groups the decision variables into different subgroups. In the optimization stage, KIDD dynamically decomposes the subgroups whose dimension is still high based on the decomposition information of HPDG, as described in Algorithm 2. The CC proposed in this paper can decompose different types of problems before and during the optimization stage, whereas the traditional CC decomposes the problems only before or during the optimization.

(1)Initialize population;
(2)subgroupsHPDG (FuncID, D);
(3)for FEs Maximum do
(4) subgroupsKIDD (subgroups);
(5) Traditional CC or contribution-based CC optimizes subgroups;
(6)end for
(7)return fitness;
3.2. Hybrid-Pool Differential Grouping
3.2.1. Unit Vector-Based Perturbation

To determine whether two sets of variables are related, we apply perturbation to the sets. According to the IEEE 754 arithmetic [40], a large number of floating-point numbers affect the computational accuracy, and the HPD proposed in this paper uses the fixed-value perturbation described by equations (6) and (7) which affects the identification accuracy. We consider two types of problems divided in [41]:

Type 1: The problems whose fitness value changes a lot even when very small upper bounds and very great lower bounds are set to those decision variables.

Type 2: The problems whose fitness value changes slightly when very great upper bounds and very small lower bounds are set to those decision variables.

The two types of changes in fitness values listed above are absolute changes, and relative change of fitness values is defined as follows.

Definition 1. Let the differences between the maximum and minimum values of problem 1 and problem 2 be and , and the changes in fitness values are and for the same perturbation done for both problems, respectively. If , then the change in fitness values of problem 1 is greater than that of problem 2.
Definition 1 means that the magnitude of the change in fitness is compared relatively. The perturbation is applied to the set of variables according to the following equation. According to equation (5), the UVP strategy is described in equations (8) and (9), where and are real numbers greater than 0.Subsets and of decision vector are initialized. Then, the unit vectors and of and are calculated separately, and the variables in the decision vectors and that do not belong to are assigned zero, and then the perturbations are applied to the currently identified subsets of variables, respectively.

3.2.2. Hybrid-Pool Decomposition Framework

The hybrid-pool decomposition framework is shown in Figure 2, where the detection pool and update pool are merged, all known groupings as detection pool and update pool and unidentified variables as another detection pool with one unidentified subgroup. The decomposition produce of HPDS is as follows. The current variable identifies relationships with other variables in detection pool and update pool, as shown in Figure 2. The identification between variables is divided into two parts: one is the identification of with the unidentified variables set V, and the other is the identification of with each identified subgroup through binary interaction, as shown in lines 5-6 of Algorithm 3; the interaction details are shown in Algorithm 4. Then the results of the interaction between two parts are combined as the steps in lines 7–17 of Algorithm 3. Zero means that no variable set is related to , and nonzero number means the index of the set related to .

(1)
(2)set seq and subgroups to ;
(3)for do
(4)
(5)
(6)
(7)ifthen
(8)  insert into ;
(9)end if
(10)  ifthen
(11)   create new ;
(12)   insert into ;
(13)  end if
(14)  ifthen
(15)   insert into ;
(16)  end if
(17)  delete from V;
(18)end for
(19)return subgroups;
(1)ifthen
(2)
(3)
(4)
(5)
(6)
(7)
(8)ifthen
(9)  ifthen
(10)   return middle;
(11)    else
(12)    
(13)  end if
(14)else
(15)   ifthen
(16)    return 0;
(17)   else
(18)   
(19)  end if
(20)end if
(21)end if

The steps of binary detection for the identification of the current variable with other variables are described in Algorithm 3. First, interacts with all known subgroups at one time; if is not related to all subgroups, algorithm exits; otherwise, the known subgroups will be divided into two equal-sized parts, and interacts with one part of the known subgroups, and if they are related, the subgroups continue recursion; otherwise, interacts with the other known subgroups recursively. Continue the above operations until the subgroup related to is identified.

3.2.3. Decomposition Complexity Analysis of HPDG

The detection complexity of the variable with known subgroups is shown in Figure 3. Let there be m subgroups after the problem is decomposed, and the detection depth of is p. The part in curly bracket indicates the set of subgroups interacting with . The number of subgroups in the th level can be expressed as , so the detection depth , and the decomposition complexity of HPDG of n-dimensional problem is .

3.3. Known Information-Based Dynamic Decomposition Method

In this section, we propose the KIDD method for the limitation that the static decomposition method cannot decompose fully nonseparable problems. KIDD and HPDG together constitute the TSD method to improve the performance of CC solving LSGO problems. KIDD is detailed in Algorithm 5.

(1)whiledo
(2)  ifthen
(3)  random decomposition
(4)  update subgroups;
(5)end if
(6)end while
(7)return subgroups;

KIDD checks all subgroups’ dimension decomposed by HPDG; if the subgroup’s dimension exceeds subD, the subgroup is decomposed by random grouping [18] and update subgroups set. subD denotes subgroup dimension, and the value of subD is set according to the performance of the optimizer used by CC.

4. Experiment Results and Discussion

4.1. Experimental Methodology

The performances of the decomposition methods were compared on the IEEE CEC’2010 and CEC’2013 special sessions on large-scale global optimization [42, 43]. The experiments in this paper were divided into two parts. In the first part, HPDG was compared with DG [24], DG2 [25], RDG [27], RDG2 [28], and ERDG [30] on problem decomposition. HPDG set as RDG did, and the values of α used to calculate are shown in Tables 1 and 2.

Three evaluation metrics were selected to evaluate the performance of the decomposition methods:(1)The decomposition accuracy is calculated as [30]; decision variables that are corelated to each other are grouped into the same subgroup and the variables that are not corelated to each other are grouped into different subgroups. In the experimental tables, “separable” and “non-sep” columns represent the decomposition accuracy of separable and nonseparable variables, respectively.(2)The number of function evaluations (FEs) is used to decompose the problem [42, 43].(3)In this paper, the growth rate of FEs is introduced to evaluate the performance of decomposition methods on high-dimensional problems, as in the following equation:where denotes the growth rate of FEs, and and denote the FEs of two different dimensional problems, respectively.

In the second part, TSD and other decomposition methods embedded in DECC were compared on problem optimization. Adaptive differential evolution with optional external archive (JADE) [44] was used to optimize the problem in optimization stage. FEs are set to 3e+ 6 as the termination condition for each run of the experiment [43]. The parameters of the decomposition methods and JADE were set as in their publications.

4.2. Decomposition Comparison
4.2.1. Comparison on Perturbation Strategy

In this section, we verify the effectiveness of UVP. As described in Table 3, HPDS with UVP correctly decomposes the four test functions f3, f6, f7, and f10 and improves the decomposition accuracy on f8 and f11 compared with HPDS with fixed-value perturbation (FVP). That is because UVP applies perturbation of appropriate size when two sets of the variables interact, while the perturbation applied by FVP cannot satisfy the requirements of different problems.

4.2.2. Decomposition Results of HPDG and Other Decomposition Methods

Table 4 summarizes the decomposition results on CEC’2013 special sessions on large-scale global optimization. HPDG correctly decomposes all functions except f5, f8, f11, f13, and f15. HPDG suffers from a flaw when detecting the indirect relationship between variables; if the shared variables are not grouped firstly, the indirect related variables cannot be grouped into the same subgroup. Although the number of functions correctly decomposed by HPDG is similar compared to DG2, RDG, and RDG2, HPDG uses fewer Fes; as shown in the last row of Table 3, HPDG correctly decomposes 7 functions using fewer FEs than other methods. HPDG uses the fewest FEs on 9 out of 15 functions; this is because HPDG has lower decomposition complexity compared to other methods.

4.2.3. Decomposition Results on High-Dimensional Functions

Since the structures of CEC’2013 special sessions on large-scale global optimization are fixed, the robustness of the decomposition methods is compared on extended high-dimensional CEC’2010 special sessions on large-scale global optimization. The extended high-dimensional CEC’2010 functions as the consistent set in the original CEC’2010 special sessions on large-scale global optimization [30].

Tables 59 summarize the decomposition results of the decomposition methods on 1000- to 5000-dimensional functions. Except for DG, each decomposition method has similar decomposition accuracy on the extended high-dimensional CEC’2010 functions, and all can decompose most of them correctly, but the FEs are significantly different. As the dimension grows, FEs used by DG and DG2 increase the most, RDG and RDG2 follow, and ERDG and HPDG are the best. That is because the complexity of DG and DG2 is [24, 25], and the complexity of RDG and RDG2 is [27, 28]. Table 10 summarizes the growth rates of FEs used by RDG, ERDG, and HPDG, which are calculated by equation (10). The least growth of FEs is used by ERDG and HPDG; this is because ERDG uses the historical information of variable interactions to avoid examining some interrelationship between variables [30], and the complexity of HPDG is , which is the lowest. The FEs used by HPDG and ERDG are comparable. As described in Table 10, the growth rate of FEs used by HPDG is slightly better than that of ERDG.

4.3. Optimization Results of CC with Decomposition Methods

Table 11 summarizes the optimization results of the decomposition methods being embedded in DECC [24]. The DECC-TSD proposed in this paper achieves the best optimization results on six functions. Compared with the state-of-the-art DECC-ERDG, DECC-TSD is more advantageous on the fully nonseparable function f15. Therefore, DECC-TSD alleviates the dilemma of CC with static decomposition in solving the fully nonseparable problems.

On fully separable functions f1 and f2, each decomposition method decomposes the problems using the same FEs, so they achieve the same results. Because HPDG uses the least FEs for decomposing f3 and f6, both HPDG and TSD achieve the best scores. On the partly separable function f13, RDG and RDG2 achieve the best scores, because HPDG and ERDG do not decompose it correctly. The optimization results of DECC-TSD with grey background in Table 11 indicate better optimization grades compared to DECC-HPDG. In the first stage of TSD, HPDG does not reduce the dimension of the fully nonseparable function f15, and, in the second stage of TSD, KIDD detects that the dimension of f15 is high, so KIDD dynamically decomposes f15, and thus the optimization result of DECC-TSD is better than that of DECC-HPDG on f15. Similarly, the optimization result of DECC-TSD on f11 is better than that of DECC-HPDG. Although HPDG does not decompose f12 and f13 correctly, KIDD performs an effective dimension reduction of the subgroups in the optimization stage, so DECC-TSD has better grades.

5. Conclusions

In this paper, we propose a TSD to improve the performance of CC in solving LSGO problems. In the first stage of TSD, HPDG is proposed, which provides a variable detection structure HPDS with very low complexity to decompose problems, and the UVP is proposed for HPDS to improve the decomposition accuracy. HPDG further reduces the FEs used to decompose the LSGO problems. In the second stage of TSD, KIDD decomposes the subgroups with high dimension based on the decomposition information of HPDG. TSD alleviates the dilemma of CC with static decomposition in solving fully nonseparable problems and reducing the computational cost of decomposition methods.

In future work, we plan to further improve the decomposition accuracy of HPDG and research new algorithms to solve fully nonseparable LSGO problems.

Data Availability

The algorithms proposed in this paper are validated on two test suites with benchmark functions described in literatures [42] and [43].

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Acknowledgments

This work was supported by the National Natural Science Foundation of China (Grants nos. 61763002 and 62072124), Guangxi Major Projects of Science and Technology (Grant no. 2020AA21077021), and Foundation of Guangxi Experiment Center of Information Science (Grant no. KF1401).