Research Article  Open Access
Meiji Cui, Li Li, Miaojing Shi, "A Selective BiogeographyBased Optimizer Considering Resource Allocation for LargeScale Global Optimization", Computational Intelligence and Neuroscience, vol. 2019, Article ID 1240162, 17 pages, 2019. https://doi.org/10.1155/2019/1240162
A Selective BiogeographyBased Optimizer Considering Resource Allocation for LargeScale Global Optimization
Abstract
Biogeographybased optimization (BBO), a recent proposed metaheuristic algorithm, has been successfully applied to many optimization problems due to its simplicity and efficiency. However, BBO is sensitive to the curse of dimensionality; its performance degrades rapidly as the dimensionality of the search space increases. In this paper, a selective migration operator is proposed to scale up the performance of BBO and we name it selective BBO (SBBO). The differential migration operator is selected heuristically to explore the global area as far as possible whist the normal distributed migration operator is chosen to exploit the local area. By the means of heuristic selection, an appropriate migration operator can be used to search the global optimum efficiently. Moreover, the strategy of cooperative coevolution (CC) is adopted to solve largescale global optimization problems (LSOPs). To deal with subgroup imbalance contribution to the whole solution in the context of CC, a more efficient computing resource allocation is proposed. Extensive experiments are conducted on the CEC 2010 benchmark suite for largescale global optimization, and the results show the effectiveness and efficiency of SBBO compared with BBO variants and other representative algorithms for LSOPs. Also, the results confirm that the proposed computing resource allocation is vital to the largescale optimization within the limited computation budget.
1. Introduction
Evolutionary algorithms (EAs) are efficient tools to solve complex optimization problems. Biogeographybased optimization (BBO) [1], proposed by Simon in 2008, is inspired by biogeography regarding the migration of species between different habitats, as well as the evolution and extinction of species. Assuming an optimization problem and some candidate solutions, each habitat represents a candidate solution, the suitability of the habitat is the fitness of the optimization problem, and the habitat features represent decision variables. According to the biogeography theory, a superior solution tends to share more promising information with the inferior one by the way of migration, specifically high emigration as well as low immigration in this case, and vice visa. Also, mutation may occur with certain probability in accordance with the biogeography evolution.
As a new yet promising EA, BBO has been applied to solve singleobjective problems [2], multiobjective problems [3, 4], and constrained problems [5] to some success. What’s more, some extensions of BBO have been proposed to improve its performance [6, 7]. BBO has been extensively explored to deal with many realword complex problems, such as manufacturing system scheduling [8], supply chain design optimization [9], and hub competitive location [10]. However, it has been reported that the performance of BBO degraded rapidly when the problem dimension increases [11]. With the advent of big data era, the scalability of an EA is a significant indicator to be considered.
In comparison with traditional optimization problems, modern optimization problems [12, 13] tend to involve a large number of decision variables, which is also conceptualized as largescale optimization problems (LSOPs). Owing to the explosion of search space and interdependencies among decision variables, LSOPs cannot be tackled in reasonable time by conventional EAs. This has made LSOPs an open and challenging problem, which has attracted intensive attention in recent trends.
Existing methods to deal with LSOPs can be divided into two categories, i.e., decomposition methods and nondecomposition methods. Nondecomposition methods refer to those exploring some special operators [14], local search [15], and hybrid algorithms [16], etc. to improve the search ability of conventional EAs. While decomposition methods, also known as divide and conquer (DC), take advantages of the modularity characteristic of optimization problems and divide the highdimensional problem into several lowdimensional subproblems. These subproblems can thus be evolved with a certain EA independently in a more efficient manner. Due to the dimensionality mismatch brought by DC, which implies that the subsolution cannot be evaluated by the original objective function directly, it is a natural way to complement the subsolution to be evaluated as a complete solution by the combination of the representative of each subproblem, also known as cooperative coevolution (CC).
Compared with nondecomposition methods, the DC framework is more efficient and therefore more popular. Recent works along this line mainly focus on the grouping strategy for subproblem division, e.g., random grouping [17] and recursive differential grouping [18]; on the other hand, the performance of optimizers and the allocation of computing resources among subproblems within limited computational budget are also crucial but have not been largely explored yet. Therefore, it is meaningful to investigate new algorithms for LSOPs with the aim of making a new attempt for this difficult problem as well as exploring extensions of BBO.
In this paper, we intend to scale up the performance of BBO to solve the LSOPs. We propose a novel Selective Migration Operator (SMO) to balance exploration and exploitation. If the selected emigration individual is better than the immigration one, once the migration occurs, a differential migration operator with a relatively large value is chosen to share more good information with the poor individual; otherwise, a normal distributed random value with small variance is applied for local search. Through the selective migration operator, a more rapid and efficient search process can be conducted in reasonable time. Furthermore, the DC framework is adopted to enhance the ability to solve highdimensional problems. To solve the problem of subgroup contribution imbalance in the context of DC, a simple and efficient computing resource allocation strategy is proposed in the end.
The paper is set as follows. In Section 2, the BBO algorithm and LargeScale Optimization (LSO) are briefly introduced. Section 3 presents our Selective BiogeographyBased Optimization (SBBO) with selective migration operator and a more efficient computing resource allocation strategy for DC framework. Section 4 depicts the experiments and corresponding results, followed by some analysis. Finally, conclusion and future work are drawn in Section 5.
2. Background
2.1. BiogeographyBased Optimization
In biogeography, there are two important terms, namely, habitat suitability index (HSI) and suitability index variables (SIVs) [1]. HSI is used to evaluate the living environment for each habitat while SIVs are the influencing factors of HSI. For an optimization problem, the population, i.e., habitats, represents a set of candidate solutions, while the SIVs of habitats are considered as the feature representations of the candidate solutions. Therefore, the evolutionary algorithm inspired by biogeography, i.e., biogeographybased optimization, is naturally used to solve different kinds of optimization problems.
There are two main operators in canonical BBO, i.e., migration operator and mutation operator. The migration operator is to share search information among individuals, and the mutation operator is to enhance the population diversity. The immigration rate λ_{i} and emigration rate μ_{i} of a habitat H_{i} can be calculated by the migration model, which is shown in Figure 1 [1]. More specifically, we adopt a simplified linear migration model to demonstrate the process, where the migration model is the function of the number of species. When the number of species increases, fewer species can survive for immigration and more species tend to emigrate to other habitats, and vice visa. The corresponding immigration and emigration rates are given bywhere I is the maximum immigration rate, E is the maximum emigration rate, S_{i} is the number of species of the habitat H_{i}, and S_{max} is the maximum number of species. In BBO, the habitat with more species signifies a better solution. That being said, a better solution has lower immigration rate and higher emigration rate, so that it can share promising information with other solutions and is less likely to be destroyed due to migration.
Next, the migration can be expressed aswhere H_{i} is the immigration habitat and H_{j} is the selected emigration habitat. SIV is a suitability index variable which represents the feature of the habitat. Equation (3) means that the SIV of the habitat H_{i} can be replaced by the SIV of the selected habitat H_{j}.
Mutation operator is a probabilistic one that can modify solution features, which is like mutation in many other EAs [19]. The purpose of mutation is to increase diversity among the population. The pseudocode of the canonical BBO is described in Algorithm 1.

Extensive works have been analyzed and discussed since BBO was proposed. With respect to different migration models corresponding to nature migration phenomenon, Ma [20] proposed six different migration models, among which sinusoidal migration curves perform the best. Additionally, some efficient migration operators and mutation operators have also been proposed to improve the performance of original BBO. Ma and Simon [5] proposed BBO with blended operator to solve constrained optimization problems. Guo et al. [7] further proposed the uniform version of extended migration operator (UEMO) to enlarge the space for offspring, thus avoiding local optimum to some extent. Zhang et al. [2] merged a differential mutation operator and a sharing operator into BBO’s migration operator to balance the global and local search ability. Mi et al. [21] combined differential evolution mutation operators with simulated binary crosser of genetic algorithms. Apart from the above, some useful strategies borrowed from EAs have been applied to BBO. Gong et al. [22] combined differential evolution and BBO for numerical optimization. Zhang et al. [6] proposed a novel hybrid algorithm based on BBO and grey wolf optimizer to make full use of the two algorithms’ search ability. Khademi et al. [23] took advantages of the featuresharing capability of invasive weed optimization to enhance the performance of BBO. Lohokare et al. [24] accelerated BBO by adopting neighborhood search. To enhance the population diversity in BBO, oppositionbased learning [25] and chaos strategy [25] have been introduced. Some theoretical studies of BBO can be found in [7, 26, 27].
Due to the simplicity and efficiency, BBO has been widely adopted in many engineering and science tasks. Bhattacharya and Chattopadhyay [28] solved both convex and nonconvex economic load dispatch problems of thermal plants with the assistance of BBO. Rahmati and Zandieh [29] developed an improved BBO to deal with flexible job shop scheduling problem. Niknamfar et al. [10] took advantage of BBO to handle a new hubandcenter transportation network problem. For further interest, readers can refer to some comprehensive reviews of BBO in [30, 31].
BBO in general performs well for most lowdimensional optimization problems; notwithstanding, its performance deteriorates rapidly when it comes to the highdimensional problems. Unlike other optimization algorithms [17, 32, 33], few works on BBO aimed to scale up its performance. To the best of our knowledge, Guo et al. [7] made the first attempt to test their improved BBO with UEMO on largescale optimization problems. However, UEMO does not outperform or cannot be even compared to the stateoftheart largescale algorithms. UEMO is the first attempt to handle LSOPs, but not yet scalable for LSOPs. With the advent of big data era, more and more optimization problems tend to involve thousands or even millions of decision variables. The scalable ability of EAs is crucial to deal with modern optimization problems. Therefore, in this work, we intend to scale up the performance of BBO.
2.2. LargeScale Optimization
Largescale optimization refers to the optimization problems with large numbers of decision variables. Although there is no formal definition of LSOPs, it is typically referred to the optimization problems in the highdimensional space where conventional algorithms [17] suffer from the “curse of dimensionality” and fail to locate the optimum. Three reasons account for the failure: (1) with an increase of the decision variables, the corresponding search space will exponentially increase, which makes it difficult to optimize searching in such large space; (2) the characteristic of problem may be altered due to the increase of dimensionality; (3) evaluating LSOPs is timeconsuming and sometimes unrealistic for realworld optimization problems which require to be solved in a reasonable time. Over the last decade, plenty of works have been proposed to copy with LSOPs. Basically, they can be divided into two categories: decomposition methods and nondecomposition methods.
2.2.1. Decomposition Algorithms
Decomposition methods adopt the strategy of divide and conquer. It contains two steps, namely, decomposition and optimization. In the decomposition stage, a highdimensional problem is decomposed into several lowdimensional subproblems which are easier to handle. In the optimization stage, each subproblem is evolved independently using one or several EAs. The final solution is a concatenation of representatives from each of the subproblem. Three crucial issues should be considered in this procedure, i.e., the decomposition accuracy, selection of optimizer, and computing resource allocation to the subcomponents.
The purpose of decomposition is to divide the interacting variables into a subcomponent, such that the global optimum can be obtained by evolving each lowdimensional subproblem independently. Early decomposition methods [17, 34, 35] does not explore variable interactions, thus failing to handle nonseparable problems. Recently, many research works have started to address this issue by implicitly or explicitly detecting the variable interactions. Sun et al. [36] proposed a statistical variable interdependence learning (SL) scheme based on nonmonotonic detection to explore variable interdependence. Omidvar et al. [37] proposed a differential grouping (DG) method based on nonlinear detection. To enhance the accuracy and efficiency of decomposition, some improved methods were proposed, such as extended DG (XDG) [38], DG2 [39], and recursive DG (RDG) [18].
Potter and De Jong [40] initially applied DC framework to improve the performance of GA. Since then, many metaheuristic algorithms, e.g., differential evolution [17], particle swarm optimization [34], and artificial bee colony [41], have demonstrated their superiorities in solving the LSOPs in the context of DC. Nevertheless, few works have focused on the scalability of some new yet efficient EAs, while in our study, we specifically scale up BBO to deal with LSOPs.
In the original DC framework, each subgroup is evolved in a roundrobin fashion with equal computational budget allocated. It has been reported that the contribution of each subgroup to the global fitness of the individuals was in fact varied [42]. Omidvar et al. [42] proposed a contributionbased cooperative coevolution that selects the subgroup to be evolved according to their contributions to the global fitness. The contribution was calculated accumulatively, which can be greatly favored from the components with a good initial contribution. It cannot respond to the timely change of objective value in particular in the late phase of evolution. Therefore, Omidvar et al. [43] mended the contribution calculation formula later. Yang et al. [44] instead proposed to discard the stagnant components if detected so that the limited computing resource can be saved. Nevertheless, they might also remove the components that could be temporal stagnant. Different from above studies in serial computing environment, Jia et al. [45] proposed an adaptive resource allocation scheme in the distributed computing environment. Compared to other issues in the DC framework, computing resource allocation of subgroups has been paid less attention, however, which is closely related to practical application.
2.2.2. Nondecomposition Algorithms
In addition to the CC, another research line to address the LSOPs is to improve the performance of traditional algorithms. Representative techniques include efficient initialization methods [46]; special operators for sampling and mutations [47, 48]; and hybrid algorithms [16] to accumulate strengths of different algorithms. To reduce the computation cost, surrogate model [49–51], and parallel computing [52, 53] have also been investigated to solve LSOPs.
Overall, it is meaningful to scale up the performance of BBO with the strategy of cooperative coevolution to deal with LSOPs in the big data era. Although DC has been embedded into canonical BBO, i.e., CBBO, it was only tested on functions of 30 dimensions [54]. The performance of CBBO on highdimensional problems (larger than 100 dimensions) is still unknown. Hence, we propose a selective migration operator to balance the ability of exploration and exploitation; the DC framework is utilized as well where we introduce a more efficient strategy to allocate the limited computational budget.
3. Proposed Approach
3.1. Selective Migration Operator
A Heuristic Migration Operator (HMO) was proposed in reference [7]. Assuming that H_{j}(SIV) is selected to immigrate from H_{i}(SIV), if the fitness of H_{j}(SIV) is better than that of H_{i}(SIV), then H_{j}(SIV) will share good information with H_{i}(SIV) by migration. Otherwise, the migration will not happen. The heuristic migration operator can be represented as follows:where , f is the fitness value (we consider the minimization problem in our paper, unless otherwise specified). What’s more, they extend the value of to enlarge the search area, which is called Uniform version of Extended Migration Operator (UEMO). In HMO and UEMO, the good emigrated individual intends to share promising information with the poor one, while the poor emigrated individual will not influence the good one. However, the current good individual will not be evolved in this generation, which degrades the exploitation ability. What’s more, the global optimum is more likely to be located around these good individuals. Therefore, we design a Selective Migration Operator (SMO) to enhance the exploitation ability.
To accelerate the convergence of the local search with better accuracy, we propose a normal distributed migration operator. The normal distribution curves with various standard deviations are shown in Figure 2. Since we focus on local search, smaller variations are preferred. Inspired by the HMO, we propose a Selective Migration Operator (SMO) to balance the exploration and exploitation. The selective migration operator can be represented as follows:where β is a variable close to 1, and γ is a normal distributed random number with smaller variations. In SMO, the poor immigrated individual will learn more useful information from good emigrated one, while the good immigrated individual will exploit its neighborhood area. The pseudocodes of SMO are given in Algorithm 2. Since the individuals in BBO are mutated towards random direction through mutation operator which may destroy good individuals, the mutation operator was removed. We use the selective migration operator to replace the original migration operator and name the corresponding algorithm selective biogeographybased optimization (SBBO).

3.2. Resource Allocation Based on Contribution
Since cooperative coevolution scheme is efficient for highdimensional problems, we adopt DC framework for LSOPs in our paper. As we discussed above, it is unwise to assign equal computational budget to each subgroup due to the imbalanced contribution of them to the global fitness value. To address this issue, a contributionbased resource allocation scheme needs to be considered, which yields the essential question about how to measure each subgroup’s contribution to the overall fitness value. The previous contribution calculation methods either focus too much on the initial good solutions [42] or brutally abandon the stagnant subgroups [44]. We instead calculate the contribution by the Relative Fitness Improvement (RFI). More specifically, the relative fitness improvement of subgroup i at generation t (generation refers to evolution of each subgroup) is defined aswhere and refers to the best overall fitness value before and after subgroup i undergoes the evolution, respectively. In the first cycle (a cycle refers to a complete evolution of all subgroups), each subgroup is evolved by sequence. The RFI values of each subgroup is calculated according to equation (7) and stored in an archive. Then, the subgroup i with largest RFI value is selected to undergo evolution in the next generation. And the RFI value of the subgroup i is updated after evolution so that RFI is in a dynamic updated manner. The pseudocodes of resource allocation based on RFI are presented in Algorithm 3.

3.3. Proposed Method
As discussed above, to deal with the LSOPs in the context of DC, we propose to use SBBO as the base optimizer and allocate the computing resource to different subcomponents according to the RFI. Nevertheless, the computing resource will still be assigned to the subgroup of extremely small RFI value in the late phase of evolution. Thereby, the improvement of the overall best fitness value is not obvious. Other subgroups considered as stagnant ones before may be promising after several evolutions. Hence, to avoid wasting the computing resource on stagnant subgroup, an extra constraint is applied. If the RFI of subgroup i is smaller than a small value, it can be regarded as a temporal stagnant one and discarded from evolutionary cycle temporarily. If all the subgroups are considered as stagnant ones, each subgroup will be evolved equally, and the RFI will be updated completely. That is to say, the extra constraint added to the resource allocation strategy can further enhance the efficiency of computing budget. We name the SBBO, in the context of CC, with the resource allocation strategy after CC_SBBO_RA, although many different decomposition strategies have been proposed. Given decomposition accuracy and computational efficiency, we adopt RDG to divide the optimization problems in this paper [18]. Instead of detecting variables interactions in a pairwise manner, RDG can reduce the time complexity of decomposition by recursively examining the interaction between a selected decision variable and the remaining variables, such that more computational resource can be focused on the optimization stage. The pseudocodes of CC_SBBO_RA are shown in Algorithm 4.

4. Experiments
Experiments consist of three parts. First, some parameters need to be determined in CC_SBBO_RA. Hence, parameter sensitivity is analyzed in the first part. Second, the SBBO algorithm with DC framework is evaluated on CEC 2010 benchmark suite. BBO variants, SaNSDE [17], and CMAES [55] for LSOPs are compared with SBBO in terms of solution accuracy, since SaNSDE and CMAES are used in the context of CC, named as CC_SaNSDE CCCMAES in the paper. In the third part, we provide the study of the contributionbased resource allocation in DC framework to show its effectiveness for LSOPs.
4.1. Benchmark Functions and Experimental Settings
The functions selected to evaluate the algorithm in our paper are CEC 2010 benchmark suite for LSGO [56]. Almost all LSO algorithms were evaluated on this benchmark suite. The CEC 2010 benchmark consists of 20 functions, listed in Table 1.
 
Note. m is the group size, and D is the dimension. In the CEC’2010 benchmark suite, m = 50, D = 1000. 
4.2. Parameter Sensitivity
In the proposed method, three parameters need to be determined before the experiment. In SBBO, β, a learning constant, determines how much information will be shared between the individuals. To investigate the constant β, we examine the change of fitness on both uni and multimodal test problems with varying degrees of separability (f_{4}, f_{5}, f_{9}, and f_{10} from Table 1). The fitness averaged over 25 independent runs as β increases is shown in Figure 3, from which we note that the fitness profiles on both uni and multimodal problems with varying degrees of separability are a bit different. It is straightforward that β = 0.9 performs best. Through the fitness comparison, as we discussed above, only better individual’s information can be emigrated to the evolved individual. As we all know, more good information sharing can result in faster convergence. Therefore, a large constant (close to 1) is preferred, which is confirmed in the experiments. When β = 0.5 or β = 0.7, only a relative small part of promising features can be shared, which degrades the information communication between individuals to some extent. When β is larger than 1, more uncertain information will be introduced to deteriorate the evolved individual. Hence, β = 0.9 is adopted here.
(a)
(b)
(c)
(d)
In BBO, γ is a normal distributed random number with smaller variations, which determines the local search ability. To investigate the appropriate variation, the same setting except the variation (0.1, 0.2, and 0.3), the change of fitness is shown in Figure 4. It is obvious that γ = norm (0 and 0.2) performs best except f_{10}, which is a multimodal function. If the variation is 0.1, the local area is too small to search. While the variation is 0.3, the local search is too large so that it cannot be exploited enough. In this paper, γ = norm (0 and 0.2) is adopted.
(a)
(b)
(c)
(d)
In the CC_SBBO_RA, the threshold value ξ, an extra constraint that determines which subgroup is in the temporal stagnation condition, needs to be explored in detail. As discussed above, RFI is used to measure each subgroup contribution, based on which the subgroup to be evolved is selected. That is to say, the smaller the RFI, the more likely the related subgroup to be stagnant. Since RFI is a relative value, we observe the change of fitness over different ξ values (0.1, 0.01, 0.015, and 0.001). When ξ is a large value (such as 0.1), as shown in Figure 5, the constraint will be too strict to determine stagnation. When ξ is too small, limited computing resource will be still assigned to stagnate subgroups. From the empirical experiment, ξ = 0.015 performs best, which is adopted in the paper.
(a)
(b)
(c)
(d)
4.3. Comparisons of BBO with Its Variants and Other Representative Algorithms
To the best of our knowledge, UEMO [7] was the first attempt to evaluate BBO variant’s performance on LSOP benchmarks. UEMO adopted an extended migration operator to avoid the issue of shrinking the searching space due to blended migration operator. UEMO outperformed the original BBO w.r.t both best and average performance for LSOPs. As the best BBO variant for LSOPs, we compare our SBBO with it. Both algorithms are embedded into DC framework with the strategy of cooperative coevolution, every algorithm is called CC_Algorithm. The decomposition method adopted in our paper is RDG [18], which is the most accurate and efficient method so far. The total fitness evaluations (FEs) is 3e6 both for decomposition and optimization.
The best, mean, standard deviation values are presented in Table 2. CC_SBBO significantly outperforms CC_BBO on all benchmark problems. Furthermore, CC_SBBO, compared with CC_UEMO, achieves best solution quality on 17 benchmark functions and is competitive for the rest 3 functions. CC_SBBO’s efficiency is attributed to the fact that the selective migration operator keeps the good exploration ability and focuses more on exploitation compared to the other migration operators.
 
Note. The notation “↑/‖/↓” represents that CC_SBBO generated statistically “better/equallywell/worse” solution than the other algorithms. The best performances are highlighted bold. 
SaNSDE [17], as a base optimizer, is widely used to solve LSOPs due to its efficiency, which adopts the strategy of neighborhood search and adaptation [57]. As an efficient and most used EA for LSOPs, CC_SaNSDE is compared with CC_SBBO, as shown in Table 2. CC_SBBO performs better than CC_SaNSDE on 5 benchmark functions, especially for fully separable functions. CC_SBBO can compete with CC_SaNSDE on function 6, 11, 14, 15, and 19. The good performance of CC_SBBO attributes to the proposed selective migration operator which increases its global search diversity and local search ability. In addition, migrated individuals and immigrated individuals of SBBO are selected according to the migration curve with a certain probability rather than random selection, which improves its performance to some degree. CC_SaNSDE performs better than CC_SBBO on the other 10 functions due to its varied neighborhood search operators and parameter adaptation. From the statistical results, CC_SBBO cannot beat CC_SaNSDE completely but it still has some advantages over CC_SaNSDE in some aspects as we mentioned before. Although both CC_SaNSDE and CC_SBBO perform worse than CC_CMAES on most functions, SaNSDE is still widely used as a base optimizer to deal with LSOPs due to its fast convergence. Analogue to SaNSDE, CC_SBBO provides us an alternative algorithm to deal with LSOPs, especially for some fully separable problems.
As an efficient algorithm for LSOPs, covariance matrix adaptation evolution strategy (CMAES) possesses a specific sampling strategy which samples offspring through a multivariate Gaussian distribution [58]. Also, this distribution is updated according to the offspring. From Table 2, CC_CMAES achieves best results on 13 functions due to its sampling strategy. The distribution estimated from the population can represent the correlation between decision variables. Thus, it is natural that CC_CMAES performs best on most partial separable and nonseparable functions, as indicated in [58]. Moreover, CC_CMAES can achieve good performance dealing with functions of rotation characteristic, and most test functions used in the paper possess the rotation characteristic. However, the performance of CC_CMAES deteriorates when it deals with fully separable and multimodal functions, such as function 2 and 3. Since there is no correlation between decision variables, the advantage of its sampling strategy declines to some extent. Moreover, CC_CMAES is more prone to getting stuck in local optimum when dealing with largescale multimodal problems with no correlation between decision variables. We cannot ignore that some fully separable and multimodal problems do exist in the real world. In that cases, CC_SBBO can perform better than CC_CMAES according to Table 2. It is worth noting that, as pointed in [59], the initial candidate solution and the initial global step size of CMAES must be chosen problem dependent, also, the optimum should presumably be within the cube . That is to say, the parameters of CMAES need elaborate adjustment for different problems, while SBBO and SaNSDE are random initialized avoiding complex parameter tuning and are not limited to the region of the optimum. Furthermore, it is of promising potential to improve the performance of both SBBO and SaNSDE to cope with rotated functions by taking advantage of the characteristic of CMAES.
4.4. Efficiency of Resource Allocation
Contributionbased cooperative coevolution was first proposed to deal with imbalanced largescale problems [42]. Each group is measured by the accumulated contribution, which shows preference for the good initial groups. The calculated contribution for each group i at cycle t can be expressed as follows:where and refer to the best overall fitness value before and after subgroup i undergoes the evolution, respectively, and is the calculated contribution of group i at cycle t−1. In this paper, we combine the aforementioned contribution measurement method with SBBO in the context of CC as the comparison algorithm, named as CC_SBBO_CB, and take it in comparison.
To save computation resource, the subgroups are out of evolution if they are considered as stagnant ones [44]. If mean and standard deviation of individuals remain unchanged for several successive generations, this subgroup is regarded as stagnation. To weaken the importance of initial good groups, they calculated the contribution of each group i at cycle t can be expressed as follows:we consider the framework of resource allocation in the context of CC, and name it CC_SBBO_FR.
Our proposed computing resource allocation (RA) is considered both in CC_UEMO and CC_SBBO, called CC_UEMO_RA and CC_SBBO_RA correspondingly. The results are presented in Table 3, and the evolutionary process is shown in Figure 6. It can be seen from Figure 6 that our contributionbased computing resource allocation scheme can greatly enhance the convergence rate and the solution accuracy except for problems f_{10} and f_{15}, which are multimodal functions and easy to be trapped in local optimum. It is obvious that CC_SBBO_CB can trap in local optimum easily due to its preference to good initial subgroups. Compared to CC_SBBO_CB and CC_SBBO_FR, our proposed resource allocation method can react quickly to the contribution change during evolution and hence decrease the computation budget on stagnant groups. Since f_{19} and f_{20} are totally nonseparate functions, we do not consider resource allocation between subgroups on these two scenarios. Therefore, CC_SBBO_RA performs best on separable and partial separable functions. To conclude, our proposed contributionbased resource allocation scheme performs efficiently for LSOPs.
 
Note. The notation “↑/‖/↓” represents that CC_SBBO_RA generated statistically “better/equallywell/worse” solution than the other algorithms. The best performances are highlighted bold. 
(a)
(b)
(c)
(d)
(e)
(f)
(g)
(h)
(i)
(j)
(k)
(l)
(m)
(n)
(o)
(p)
(q)
(r)
5. Conclusion
In this paper, we propose a selective migration operator for BBO. The selective migration operator can enhance the exploitation ability as well as keep its good exploration ability compared with the original migration operator. When dealing with LSOPs, the cooperative coevolution framework is adopted in our paper. To address the imbalance contribution of each subgroup to the overall fitness value in the context of DC, a more efficient contributionbased resource allocation method is proposed. The relative performance improvement is utilized to measure the contribution as it reflects the recent improvements timely. Also, a threshold strategy, as an extra constraint, is adopted to measure whether the subgroup is stagnant. Computing resource will not be assigned to the stagnant subgroup in the cycle. The CEC’2010 largescale benchmark functions were used to evaluate the performance of CC_SBBO_RA. From our experimental results, several conclusions can be drawn.
Firstly, BBO with selective migration operator can significantly improve the performance for LSOPs compared with other BBO variants, especially for those fully separable problems. Secondly, our proposed contributionbased resource allocation method can clearly enhance the EAs’ performance when embedded into the DC framework.
In the future, we intend to improve the performance of BBO dealing with largescale multimodal optimization problems. Also, it is interesting to explore an adaptive value for stagnation measurements with high accuracy.
Data Availability
The data used to support the findings of this study are available from the corresponding author upon request.
Disclosure
The paper was reported in Doctoral Workshop on Application of Artificial Intelligence in Manufacturing, organized by Tongji University and Lorraine University, in June 2019.
Conflicts of Interest
The authors declare that they have no conflicts of interest.
Acknowledgments
This research has been supported by the Key Research and Development Project of National Ministry of Science and Technology under grant no. 2018YFB1305304, the National Natural Science Foundation of China under grant no. 61873191, and the International Joint Training of Interdisciplinary Innovative Talents for Postgraduates of Tongji University under grant no. 2019XKJC007.
References
 D. Simon, “Biogeographybased optimization,” IEEE Transactions on Evolutionary Computation, vol. 12, no. 6, pp. 702–713, 2008. View at: Publisher Site  Google Scholar
 X. Zhang, Q. Kang, Q. Tu, J. Cheng, and X. Wang, “Efficient and merged biogeographybased optimization algorithm for global optimization problems,” Soft Computing, vol. 23, no. 12, pp. 4483–4502, 2019. View at: Publisher Site  Google Scholar
 H. Ma, Z. Yang, P. You, and M. Fei, “Multiobjective biogeographybased optimization for dynamic economic emission load dispatch considering plugin electric vehicles charging,” Energy, vol. 135, pp. 101–111, 2017. View at: Publisher Site  Google Scholar
 W. Guo, L. Wang, and Q. Wu, “Numerical comparisons of migration models for multiobjective biogeographybased optimization,” Information Sciences, vol. 328, pp. 302–320, 2016. View at: Publisher Site  Google Scholar
 H. Ma and D. Simon, “Blended biogeographybased optimization for constrained optimization,” Engineering Applications of Artificial Intelligence, vol. 24, no. 3, pp. 517–525, 2011. View at: Publisher Site  Google Scholar
 X. Zhang, Q. Kang, J. Cheng, and X. Wang, “A novel hybrid algorithm based on biogeographybased optimization and grey wolf optimizer,” Applied Soft Computing, vol. 67, pp. 197–214, 2018. View at: Publisher Site  Google Scholar
 W. Guo, L. Wang, C. Si, Y. Zhang, H. Tian, and J. Hu, “Novel migration operators of biogeographybased optimization and Markov analysis,” Soft Computing, vol. 21, no. 22, pp. 6605–6632, 2017. View at: Publisher Site  Google Scholar
 A. P. Rifai, H.T. Nguyen, H. Aoyama, S. Z. M. Dawal, and N. A. Masruroh, “Nondominated sorting biogeographybased optimization for biobjective reentrant flexible manufacturing system scheduling,” Applied Soft Computing, vol. 62, pp. 187–202, 2018. View at: Publisher Site  Google Scholar
 G. Yang and Y. Liu, “Optimizing an equilibrium supply chain network design problem by an improved hybrid biogeography based optimization algorithm,” Applied Soft Computing, vol. 58, pp. 657–668, 2017. View at: Publisher Site  Google Scholar
 A. H. Niknamfar, S. T. A. Niaki, and S. A. A. Niaki, “Oppositionbased learning for competitive hub location: a biobjective biogeographybased optimization algorithm,” KnowledgeBased Systems, vol. 128, pp. 1–19, 2017. View at: Publisher Site  Google Scholar
 H. Ma and D. Simon, Evolutionary Computation with BiogeographyBased Optimization, John Wiley & Sons, Hoboken, NJ, USA, 2017.
 Z. Yang, B. Sendhoff, K. Tang, and X. Yao, “Target shape design optimization by evolving Bsplines with cooperative coevolution,” Applied Soft Computing, vol. 48, pp. 672–682, 2016. View at: Publisher Site  Google Scholar
 H. F. Teng, Y. Chen, W. Zeng, Y. J. Shi, and Q. H. Hu, “A dualsystem variablegrain cooperative coevolutionary algorithm: satellitemodule layout design,” IEEE Transactions on Evolutionary Computation, vol. 14, no. 3, pp. 438–455, 2010. View at: Publisher Site  Google Scholar
 M. Weber, F. Neri, and V. Tirronen, “Shuffle or update parallel differential evolution for largescale optimization,” Soft Computing, vol. 15, no. 11, pp. 2089–2107, 2011. View at: Publisher Site  Google Scholar
 D. Molina, M. Lozano, A. M. Sánchez, and F. Herrera, “Memetic algorithms based on local search chains for large scale continuous optimisation problems: MASSWChains,” Soft Computing, vol. 15, no. 11, pp. 2201–2220, 2011. View at: Publisher Site  Google Scholar
 Y. Wang, B. Li, and T. Weise, “Estimation of distribution and differential evolution cooperation for large scale economic load dispatch optimization of power systems,” Information Sciences, vol. 180, no. 12, pp. 2405–2420, 2010. View at: Publisher Site  Google Scholar
 Z. Yang, K. Tang, and X. Yao, “Large scale evolutionary optimization using cooperative coevolution,” Information Sciences, vol. 178, no. 15, pp. 2985–2999, 2008. View at: Publisher Site  Google Scholar
 Y. Sun, M. Kirley, and S. K. Halgamuge, “A recursive decomposition method for large scale continuous optimization,” IEEE Transactions on Evolutionary Computation, vol. 22, no. 5, pp. 647–661, 2018. View at: Publisher Site  Google Scholar
 D. Whitley, “A genetic algorithm tutorial,” Statistics and Computing, vol. 4, no. 2, pp. 65–85, 1994. View at: Publisher Site  Google Scholar
 H. Ma, “An analysis of the equilibrium of migration models for biogeographybased optimization,” Information Sciences, vol. 180, no. 18, pp. 3444–3464, 2010. View at: Publisher Site  Google Scholar
 Z. Mi, Y. Xu, Y. Yu, T. Zhao, B. Zhao, and L. Liu, “Hybrid biogeography based optimization for constrained numerical and engineering optimization,” Mathematical Problems in Engineering, vol. 2015, Article ID 423642, 15 pages, 2015. View at: Publisher Site  Google Scholar
 W. Gong, Z. Cai, and C. X. Ling, “DE/BBO: a hybrid differential evolution with biogeographybased optimization for global numerical optimization,” Soft Computing, vol. 15, no. 4, pp. 645–665, 2010. View at: Publisher Site  Google Scholar
 G. Khademi, H. Mohammadi, and D. Simon, “Hybrid invasive weed/biogeographybased optimization,” Engineering Applications of Artificial Intelligence, vol. 64, pp. 213–231, 2017. View at: Publisher Site  Google Scholar
 M. R. Lohokare, S. S. Pattnaik, B. K. Panigrahi, and S. Das, “Accelerated biogeographybased optimization with neighborhood search for optimization,” Applied Soft Computing, vol. 13, no. 5, pp. 2318–2342, 2013. View at: Publisher Site  Google Scholar
 M. Ergezer, D. Simon, and D. Du, “Oppositional biogeographybased optimization,” in Proceedings of the IEEE International Conference on Systems, Man and Cybernetics, pp. 1009–1014, San Antonio, TX, USA, October 2009. View at: Google Scholar
 D. Simon, M. Ergezer, D. Dawei Du, and R. Rarick, “Markov models for biogeographybased optimization,” IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics), vol. 41, no. 1, pp. 299–306, 2011. View at: Publisher Site  Google Scholar
 H. Ma, D. Simon, and M. Fei, “Statistical mechanics approximation of biogeographybased optimization,” Evolutionary Computation, vol. 24, no. 3, pp. 427–458, 2016. View at: Publisher Site  Google Scholar
 A. Bhattacharya and P. K. Chattopadhyay, “Biogeographybased optimization for different economic load dispatch problems,” IEEE Transactions on Power Systems, vol. 25, no. 2, pp. 1064–1077, 2010. View at: Publisher Site  Google Scholar
 S. H. A. Rahmati and M. Zandieh, “A new biogeographybased optimization (BBO) algorithm for the flexible job shop scheduling problem,” International Journal of Advanced Manufacturing Technology, vol. 58, no. 9–12, pp. 1115–1129, 2012. View at: Publisher Site  Google Scholar
 W. Guo, M. Chen, L. Wang, Y. Mao, and Q. Wu, “A survey of biogeographybased optimization,” Neural Computing and Applications, vol. 28, no. 8, pp. 1909–1926, 2017. View at: Publisher Site  Google Scholar
 H. Ma, D. Simon, P. Siarry, Z. Yang, and M. Fei, “Biogeographybased optimization: a 10year review,” IEEE Transactions on Emerging Topics in Computational Intelligence, vol. 1, no. 5, pp. 391–407, 2017. View at: Publisher Site  Google Scholar
 B. Akay and D. Karaboga, “Artificial bee colony algorithm for largescale problems and engineering design optimization,” Journal of Intelligent Manufacturing, vol. 23, no. 4, pp. 1001–1014, 2012. View at: Publisher Site  Google Scholar
 W. Dong, T. Chen, P. Tino, and X. Yao, “Scaling up estimation of distribution algorithms for continuous optimization,” IEEE Transactions on Evolutionary Computation, vol. 17, no. 6, pp. 797–822, 2013. View at: Publisher Site  Google Scholar
 X. Li and X. Yao, “Cooperatively coevolving particle swarms for large scale optimization,” IEEE Transactions on Evolutionary Computation, vol. 16, no. 2, pp. 210–224, 2012. View at: Google Scholar
 M. N. Omidvar, X. Li, Z. Yang, and X. Yao, “Cooperative coevolution for large scale optimization through more frequent random grouping,” in Proceedings of the IEEE Congress on Evolutionary Computation (CEC), pp. 1754–1761, Barcelona, Spain, July 2010. View at: Google Scholar
 L. Sun, S. Yoshida, X. Cheng, and Y. Liang, “A cooperative particle swarm optimizer with statistical variable interdependence learning,” Information Sciences, vol. 186, no. 1, pp. 20–39, 2012. View at: Publisher Site  Google Scholar
 M. N. Omidvar, X. Li, Y. Mei, and X. Yao, “Cooperative coevolution with differential grouping for large scale optimization,” IEEE Transactions on Evolutionary Computation, vol. 18, no. 3, pp. 378–393, 2014. View at: Publisher Site  Google Scholar
 Y. Sun, M. Kirley, and S. K. Halgamuge, “Extended differential grouping for large scale global optimization with direct and indirect variable interactions,” in Proceedings of the Conference on Genetic and Evolutionary Computation, pp. 313–320, Madrid, Spain, July 2015. View at: Google Scholar
 M. N. Omidvar, M. Yang, Y. Mei, X. Li, and X. Yao, “DG2: a faster and more accurate differential grouping for largescale blackbox optimization,” IEEE Transactions on Evolutionary Computation, vol. 21, no. 6, pp. 929–942, 2017. View at: Publisher Site  Google Scholar
 M. A. Potter and K. A. De Jong, “A cooperative coevolutionary approach to function optimization,” in Proceedings of the International Conference on Parallel Problem Solving from Nature, pp. 249–257, Jerusalem, Israel, October 1994. View at: Google Scholar
 Y. Ren and Y. Wu, “An efficient algorithm for highdimensional function optimization,” Soft Computing, vol. 17, no. 6, pp. 995–1004, 2013. View at: Publisher Site  Google Scholar
 M. N. Omidvar, X. Li, and X. Yao, “Smart use of computational resources based on contribution for cooperative coevolutionary algorithms,” in Proceedings of the 13th Annual Conference on Genetic and Evolutionary Computation, pp. 1115–1122, Dublin, Ireland, July 2011. View at: Google Scholar
 M. N. Omidvar, B. Kazimipour, X. Li, and X. Yao, “CBCC3a contributionbased cooperative coevolutionary algorithm with improved exploration/exploitation balance,” in Proceedings of the IEEE Congress on Evolutionary Computation (CEC), pp. 3541–3548, Vancouver, BC, Canada, July 2016. View at: Google Scholar
 M. Yang, M. N. Omidvar, C. Li et al., “Efficient resource allocation in cooperative coevolution for largescale global optimization,” IEEE Transactions on Evolutionary Computation, vol. 21, no. 4, pp. 493–505, 2017. View at: Publisher Site  Google Scholar
 Y. H. Jia, W. N. Chen, T. Gu et al., “Distributed cooperative coevolution with adaptive computing resource allocation for large scale optimization,” IEEE Transactions on Evolutionary Computation, vol. 23, no. 2, pp. 188–202, 2018. View at: Publisher Site  Google Scholar
 E. Segredo, B. Paechter, C. Segura, and C. I. GonzálezVila, “On the comparison of initialisation strategies in differential evolution for large scale optimisation,” Optimization Letters, vol. 12, no. 1, pp. 221–234, 2018. View at: Publisher Site  Google Scholar
 Z. Yang, K. Tang, and X. Yao, “Scalability of generalized adaptive differential evolution for largescale continuous optimization,” Soft Computing, vol. 15, no. 11, pp. 2141–2155, 2011. View at: Publisher Site  Google Scholar
 S. Tuo, J. Zhang, X. Yuan, and L. Yong, “A new differential evolution algorithm for solving multimodal optimization problems with high dimensionality,” Soft Computing, vol. 22, no. 13, pp. 4361–4388, 2018. View at: Publisher Site  Google Scholar
 R. G. Regis, “Evolutionary programming for highdimensional constrained expensive blackbox optimization using radial basis functions,” IEEE Transactions on Evolutionary Computation, vol. 18, no. 3, pp. 326–347, 2014. View at: Publisher Site  Google Scholar
 E. Li, H. Wang, and F. Ye, “Twolevel multisurrogate assisted optimization method for high dimensional nonlinear problems,” Applied Soft Computing, vol. 46, pp. 26–36, 2016. View at: Publisher Site  Google Scholar
 C. Sun, Y. Jin, R. Cheng, J. Ding, and J. Zeng, “Surrogateassisted cooperative swarm optimization of highdimensional expensive problems,” IEEE Transactions on Evolutionary Computation, vol. 21, no. 4, pp. 644–660, 2017. View at: Publisher Site  Google Scholar
 H. Wang, S. Rahnamayan, and Z. Wu, “Parallel differential evolution with selfadapting control parameters and generalized oppositionbased learning for solving highdimensional optimization problems,” Journal of Parallel and Distributed Computing, vol. 73, no. 1, pp. 62–73, 2013. View at: Publisher Site  Google Scholar
 A. Cano and C. GarcíaMartínez, “100 million dimensions largescale global optimization using distributed GPU computing,” in Proceedings of the IEEE Congress on Evolutionary Computation (CEC), pp. 3566–3573, Vancouver, BC, Canada, July 2016. View at: Google Scholar
 X.W. Zheng, D.J. Lu, X.G. Wang, and H. Liu, “A cooperative coevolutionary biogeographybased optimizer,” Applied Intelligence, vol. 43, no. 1, pp. 95–111, 2015. View at: Publisher Site  Google Scholar
 Y. Mei, M. N. Omidvar, X. Li, and X. Yao, “A competitive divideandconquer algorithm for unconstrained largescale blackbox optimization,” ACM Transactions on Mathematical Software, vol. 42, no. 2, pp. 1–24, 2016. View at: Publisher Site  Google Scholar
 K. Tang, X. Yáo, P. N. Suganthan et al., Benchmark Functions for the CEC’ 2008 Special Session and Competition on Large Scale Global Optimization, Nature Inspired Computation and Applications Laboratory, USTC, Hefei, China, 2007.
 Z. Yang, K. Tang, and X. Yao, “Selfadaptive differential evolution with neighborhood search,” in Proceedings of the IEEE Congress on Evolutionary Computation (CEC), pp. 1110–1116, Hong Kong, China, June 2008. View at: Google Scholar
 N. Hansen, “The CMA evolution strategy: a tutorial,” 2016, https://arxiv.org/abs/1604.00772. View at: Google Scholar
 C. Igel, N. Hansen, and S. Roth, “Covariance matrix adaptation for multiobjective optimization,” Evolutionary Computation, vol. 15, no. 1, pp. 1–28, 2007. View at: Publisher Site  Google Scholar
Copyright
Copyright © 2019 Meiji Cui et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.