Computational Intelligence and Metaheuristic Algorithms with Applications
View this Special IssueResearch Article  Open Access
Kaifeng Yang, Li Mu, Dongdong Yang, Feng Zou, Lei Wang, Qiaoyong Jiang, "Multiobjective Memetic Estimation of Distribution Algorithm Based on an Incremental Tournament Local Searcher", The Scientific World Journal, vol. 2014, Article ID 836272, 21 pages, 2014. https://doi.org/10.1155/2014/836272
Multiobjective Memetic Estimation of Distribution Algorithm Based on an Incremental Tournament Local Searcher
Abstract
A novel hybrid multiobjective algorithm is presented in this paper, which combines a new multiobjective estimation of distribution algorithm, an efficient local searcher and εdominance. Besides, two multiobjective problems with variable linkages strictly based on manifold distribution are proposed. The Pareto set to the continuous multiobjective optimization problems, in the decision space, is a piecewise lowdimensional continuous manifold. The regularity by the manifold features just build probability distribution model by globally statistical information from the population, yet, the efficiency of promising individuals is not well exploited, which is not beneficial to search and optimization process. Hereby, an incremental tournament local searcher is designed to exploit local information efficiently and accelerate convergence to the true Paretooptimal front. Besides, since εdominance is a strategy that can make multiobjective algorithm gain well distributed solutions and has low computational complexity, εdominance and the incremental tournament local searcher are combined here. The novel memetic multiobjective estimation of distribution algorithm, MMEDA, was proposed accordingly. The algorithm is validated by experiment on twentytwo test problems with and without variable linkages of diverse complexities. Compared with three stateoftheart multiobjective optimization algorithms, our algorithm achieves comparable results in terms of convergence and diversity metrics.
1. Introduction
Multiobjective optimization usually involves many conflicting, incomparable, and noncommensurable objectives. During the past two decades, multiobjective evolutionary algorithms (MOEAs) have obtained much more interest among optimization community mainly because of the fact that they can be suitably applied to deal simultaneously with a set of possible solutions. A number of evolutionary algorithms have been developed for multiobjective problems, such as strength Pareto evolutionary algorithm (SPEA) [1], Pareto archived evolution strategy (PAES) [2], Pareto envelopebased selection algorithm (PESA) [3, 4], microGA [5], nondominated sorting genetic algorithm II (NSGAII) [6], strength Pareto evolutionary algorithm 2 (SPEA2) [7], multiple objectives with particle swarm optimization (MOPSO) [8], nondominated neighbor immune algorithm (NNIA) [9], and MOEA with adaptive weight adjustment (MOEA/DAWA) [10]. Following the recent review of evolutionary multiobjective optimization fields [11, 12], NSGAII and SPEA2 can be considered as two representatives of stateoftheart MOEAs in current multiobjective optimization community.
The current MOEAs research mainly focuses on the following several highly related issues, such as fitness assignment, diversity maintenance, external population, combination of MOEA and local search, new dominance scheme, and manyobjective problems. However, there are little fresh work done on how to generate new solutions and most current MOEAs directly adopt traditional crossover and mutation operators. Based on the works [13, 14], we can gain multiobjective problems (MOPs) with variable linkages causing trouble to MOEAs with traditional crossover and mutation operators. It seems that it is urgent to design new scheme to generate new solutions.
Estimation of distribution algorithms (EDAs) are a new computing paradigm in evolutionary computation. A posterior probability distribution model based on globally statistical information from the selected solutions is built to generate new solutions for next generation. This new class of algorithms generalizes genetic algorithms by replacing the crossover and mutation operators by learning and sampling the probability distribution of the promising individuals of population per iteration. Working in such a way, the relationships between the variables involved in the problem domain are explicitly and effectively captured and exploited. Several EDAs have been proposed for MOPs. Khan [15] proposed multiobjective BOA (mBOA) and multiobjective hierarchical BOA (mhBOA) by combining the model building and model sampling procedures of BOA [16] and hierarchical BOA (hBOA) [17] with the selection procedure of NSGAII. They compared the performance of mBOA and mhBOA with that of NSGAII on a class of bounded difficult additively separable deceptive and hierarchically deceptive functions. Bosman and Thierens [18] combined IDEAs with nondominated tournament selection and clustering, and they used clustering to split the population into subpopulation and separate models were built for each subpopulation. Laumanns and Ocenasek [19] combined mixed BOA with the selection and replacement procedures of SPEA2. The algorithm was tested on some knapsack problems, and it was shown to dominate NSGAII and SPEA2 in most instances. Zhou et al. [20] present a way, named multiobjective EDA based on decomposition, which utilizes decomposition based techniques and probabilistic model based methods, to tackle the traveling salesman problem. However, these MOEDAs do not involve how the Pareto set distributes in the decision space.
The Pareto set of a continuous MOP is a piecewise continuous ()dimensional manifold, where is the number of the objectives, which have been investigated and applied by several scholars [21, 22]. Zhang et al. [14] exploited the property explicitly and proposed regularity model based multiobjective estimation of distribution algorithm, called RMMEDA, a regularity model based EDA for continuous MOPs. They have validated that RMMEDA can effectively deal with MOPs with variable linkages and admitted that the performance of RMMEDA declines if a MOP has many local Pareto fronts [14, 23].
As Ong et al. [24, 25] suggested, in recent years, the development of hybrid genetic algorithm is one of the most significant trends in the field of metaheuristics. Methods of this kind hybridize recombination operators with local heuristics. Such a hybrid algorithm is also called a memetic algorithm. It is clearly shown that memetic algorithms have higher search ability than traditional EMO algorithms [26]. Ishibuchi and Murata [27] are the first two scholars to propose a multiobjective genetic local search algorithm. Afterward, Xu et al. [28] proposed GLSA, Jaszkiewicz [29] proposed MOGLS, and Liu et al. [30] proposed FMOPSO. As forenamed, RMMEDA extracts globally statistical information to build the probability distribution model and performs weakly in local search. If we hybridize an effective local search operator with RMMEDA, we may get balance between exploration and exploitation in the search space. For this end, an incremental tournament local searcher operator is proposed in our study, which biases solutions with high isolation value. Our algorithm is called multiobjective memetic estimation of distribution algorithm.
In order to keep wellspread Paretooptimal solutions, dominance is employed in our algorithm. The dominance does not allow two solutions with a difference less than in the th objective to be nondominated to each other, thereby allowing a good diversity to be maintained in a population. Besides, the method is quite pragmatic because it allows the user to choose a suitable depending on the desired resolution in the th objective. Deb et al. [31] have validated that the diversity metric of MOEA is slightly better than that of NSGAII based on crowding distance.
Moreover, in [14], Zhang et al. proposed ten multiobjective problems with variable linkages. However, distribution of these ten problems is linear or almost linear in decision space. Besides, RMMEDA introduces local principal component analysis to build probability distribution model, which is linear or locallinear distribution in decision space. Therefore, most of these problems may be easy for RMMEDA. For this end, two more difficulty problems are proposed in this paper, which more accord with manifold distribution in the decision space by introducing nonlinear mapping into variables.
In the remainder of the paper, we briefly mention notations and definitions in Section 2. Thereafter, in Section 3, we briefly describe RMMEDA. Section 4 presents the proposed MMEDA. Section 5 describes our proposed two problems. In Section 6, the experimental study is demonstrated. Finally, we outline the conclusions of this paper.
2. Definitions and Notations
In our paper, we consider the following continuous MOP: where , is a decision variable vector, and is a continuous search space. is the map of decision variable space to real valued objective space. The objectives in a MOP conflict each other, and no single solution can optimize all the objectives at the same time. The Pareto front/Pareto set (PF/PS) is set of all the optimal tradeoff solutions in the objective/decision space.
Definition 1 (Pareto dominance). There is a vector , which is said to dominate (denoted by ), if and only if is partially less than , which is equal to this expression: , .
Definition 2 (Pareto optimality). A point is a random optimal point and if for every and either or there is at least one such that . In other words, this definition says that is Pareto optimal if there are no feasible vectors which can decrease some criterion without causing a simultaneous increase in at least one other criterion.
3. The Framework of RMMEDA
Under some smoothness assumptions, it could be induced from the KarushKuhnTucker condition that the PS of (1) defines a piecewise continuous ()dimensional manifold, where is the number of objectives. In [14], they give us the following model to illustrate individual solutions scattered around the PS in the decision space: where is uniformly distributed over a piecewise continuous ()dimensional manifold. is an dimensional zeromean noise vector and is the number of decision variables. Figure 1 illustrates the basic idea of RMMEDA.
In [14], piecewise ()dimensional linear models are used to approximate model in (2). Local principal component analysis is applied to partition population. In RMMEDA, the number of clusters is an experimental parameter and Zhang sets it to be 5. The solutions of each cluster are used to build a statistical model by principal component analysis and we could get the parameters of each model and noise in (2). New trial solutions are sampled from each local model.
As mentioned above, offspring solutions are generated by the statistical model and how to build the model is crucial to this algorithm. In RMMEDA, they first partition population into disjoint clusters by local principal component analysis, and more details about partition can be found in [32]. For each cluster , let be its mean and its th principal component. Compute following two equations: for . Then, set Let be the th largest eigenvalue of covariance matrix of points in , and ~: From (4) and (5), we can get the model of (2) for each cluster, and new offspring solutions are sampled from the model. The flowchart of RMMEDA is illustrated in Algorithm 1.

RMMEDA is a novel and efficient multiobjective estimation of distribution algorithms and has been validated by multiobjective problems with variable linkages. However, it may fail in locating the global PF if a MOP has many local PFs [23], and it has not been tested by famous ZDT and DTLZ problems. In the next section, we proposed multiobjective memetic estimation of distribution algorithms, which is a more efficient and effective hybrid multiobjective algorithm.
4. The Proposed Method: MMEDA
4.1. Incremental Tournament Local Searcher (ITLS)
It is known that memetic algorithms, combined with EAs and local search heuristics, can be implemented to maintain a balance between exploration and exploitation in the search space. Besides, it is clearly shown by Jaszkiewicz [26] that memetic EMO algorithms have higher search ability than traditional EMO algorithms. Several memetic EMO algorithms have been proposed and show high competitive performance. In this paper, an incremental tournament local searcher is proposed and combined with RMMEDA.
Following Liu et al.’s recent reviews [30], the issues considered in the design of the local searcher operator include the selection of appropriate search direction; the selection of appropriate solutions for local optimization; and the allocation of computational budget for local search. With these notions in mind, we devise our local searcher as follows.
An active subpopulation is selected by tournament; that is, nondominated solutions are selected before modeling in RMMEDA per iteration. At first, a tournament pool is built by nondominated solutions with higher value of crowding distance, and the size of tournament pool is . If the number of nondominated solutions found so far is less than , all of the nondominated solutions are included in the tournament pool. Then we select solutions with higher value of crowding distance by performing tournament selection, and the winner solutions are put in the active subpopulation, whose size is . Since is always larger than the size of tournament pool, , we call it incremental tournament local searcher. Then, traditional genetic operators are employed on the active subpopulation for local search. Figure 2 illustrates the basic idea of ITLS.
By Figure 2, we can obtain that if a nondominated solution locates in more isolated regions, there are more offsprings created near the solutions. The aim is that the larger the crowdingdistance value of an individual, the more the times the individual will be reproduced. So there exist more chances to do search in lesscrowded regions of the tradeoff fronts. Now, we can answer the former three issues of local searcher. The search direction is optimization direction and is illustrated in Figure 2; the solutions for local optimization are biased towards isolated ones; and the allocation of computational resource for local searcher is determined by , the size of active subpopulation.
In order to present ITLS in detail, the procedure is outlined in Pseudocode 1. Before we describe the ITLS, let us give some more notations. The tournament scale is , and population is from RMMEDA before modeling. Nondominated solutions from are denoted by .

4.2. The Adopted εDominance
It is clear from the existing studies that there are two distinct goals in the development of an MOEA: the first one is convergence to the true Paretooptimal front; then the second one is maintenance of a welldistributed set of nondominated solutions. The proposed ITLS in former subsection corresponds to the first goal. In order to maintain wellspread Paretooptimal solutions, we introduce the dominance in our paper.
The dominance does not allow two solutions with a difference less than in the th objective to be nondominated to each other, thereby allowing a good diversity to be maintained in a population. By dominance, the search space is divided into a number of grids (or hyperboxes) and the diversity is maintained by ensuring that a grid or hyperbox can be occupied by only one solution. In our MMEDA, there are two coevolving populations: an internal population and an archive population . The archive population is updated based on the dominance concept, whereas a usual domination concept is used to update the internal population. Every solution in the archive is assigned an identification array , where is the total number of objectives) as follows: where is the minimum possible value of the th objective (default value of it is and is the allowable tolerance in the th objective below which two values are meaningless to the user. With the identification arrays calculated for the offspring and each archive member, we can decide how the archive population updates. More details of dominance can be found in [31].
4.3. Details of MMEDA
As mentioned above, RMMEDA extracts globally statistical information to build the probability distribution model and then samples it to generate offspring, which emphasizes global statistics ability, yet brings the trouble of weak performance in local search. However, several memetic MOEAs have been proposed and implemented to maintain a balance between exploration and exploitation in search space, which is often crucial to the success of the search and optimization process. For this end, a local search operator, called ITLS, is proposed to combine with RMMEDA. Besides, in order to maintain wellspread Paretooptimal solutions, dominance is employed in our algorithm, whose name is MMEDA. The proposed algorithm is hybrid with global search and local search, and the success of it is due to the tradeoff between the exploration ability of RMMEDA and the exploitation ability of ITLS. The details of MMEDA are in Algorithm 2.

5. Two Novel Multiobjective Problems with Variable Linkages
A number of test problems for multiobjective optimization have been designed by some scholars, such as SCH by Schaffer [33], KUR by Kursawe [34], FON by Fonseca and Fleming [35], ZDTs by Zitzler et al. [36], and DTLZs by Deb et al. [37]. These MOPs have been used in a number of significant past studies of EMO to test an optimization methodology. Following Deb’s recent review of multiobjective test problems, many of the existing test problems are separable variablewise or possess linear functions of the variable, which may not provide adequate difficulties to an EMO methodology. Zhang et al. [14] realized the point and proposed ten multiobjective problems with variable linkages (F1~F10 in Table 1).

Analyzing the ten problems proposed by [14], we can get that , , for F1~F4, and , , for F5~F10, when all the solutions converge to Pareto optimal fronts. The scheme of introducing variable linkages proposed in the ten problems can be regarded as variable linear or nearlinear mapping. Besides, RMMEDA is based on probability distributions model built by local principal component analysis, which is also a linear or locallinear statistical model. Consequently, RMMEDA may be efficient to most of these problems. In order to investigate the performance of the optimization methodology, more problems should be employed and tested. For this end, we introduce twentytwo multiobjective problems with and without variable linkages in our experiment in Table 1.
The variable linkages in our proposed problems are based on the following nonlinear mapping on the variables: In Figure 1, we can obtain that the Pareto set, in the decision space, of a continuous multiobjective optimization problem is a piecewise continuous ()dimensional manifold, and it seems that the ten problems (F1~F10) may be a simple and special case of the regularity. The problem introduced by nonlinear mapping in (7) seems more accordant with the regularity. Furthermore, there is no common rules of why we choose sin() function, and other nonlinear function can be used, and more difficulty problems can be proposed by modifying the period of the sin(). The details of two advanced problems are in Table 1 (AF1 and AF2).
6. Experimental Study
6.1. Test Problems
The first five ZDT problems were developed by Zitzler et al. [36] (so called ZDT problems), and the next five DTLZ problems were defined by Deb et al. [37] (so called DTLZ problems). These problems have been cited in a number of significant past studies in EMO community and they can test evolutionary multiobjective optimization algorithms in different aspects. Furthermore, ten problems with variable linkages proposed by Zhang et al. [14] have also been introduced in our paper, which can bring trouble to most of variablewise EMO methodology, and have been validated by Zhang et al. Lastly, two more difficult problems proposed by ourselves are also presented in Table 1.
It is necessary to note that the performance of an MOEA in tackling multiobjective constrained optimization problems may largely depend on the constrainthandling technique used [38], so we do not mention sideconstrained problems in this study.
6.2. Performance Metrics
Zitzler et al. [39] suggested that for a objective optimization problem, at least performances are needed to compare two or more solutions and an infinite number of metrics to compare two or more sets of solutions. Coello Coello et al. [8] proposed three issues to allow a quantitative assessment of the performance of a multiobjective optimization algorithm. As we know, it is a common task for any multiobjective optimization algorithm to find solutions as close as possible to the Pareto front and to make them as diverse as possible in the obtained nondominated front. Furthermore, the latter case includes maximizing the spread and the uniformity of solutions found in the final Pareto front. For this end, three metrics are employed in our paper to investigate the performance of the algorithm. As [14], we apply inverted generation distance to the final Paretooptimal set obtained by an MOEA to evaluate its convergence and spread performance. We also adopt convergence metric proposed by Deb et al. [6] to measure the convergence to the final solutions. Finally, in order to check the uniformity of Paretooptimal solutions we get in final generation and spacing devised by Schott [40] metric is employed in our paper. The three metrics are summarized as follows.
6.2.1. Inverted Generation Distance
Let be a set of uniformly distributed points in the objective space along the PF. Let be an approximation to the PF; the inverted generational distance from to is defined as where is the minimum Euclidean distance between and the points in . The inverted generation distance denotes the metric convergence and spread, which represents the distance between the set of the true Paretooptimal fronts and converged Pareto solutions obtained by EMOAs.
6.2.2. Convergence Metric
Metric is used to estimate how far the elements in the set of nondominated solutions found so far are from those in the Paretooptimal set, and it is defined as where is the minimum Euclidean distance between and the points in . Since multiobjective algorithms can be tested on problems having a known set of Paretooptimal solutions, the calculation of the two metrics is possible.
6.2.3. Spacing
Let be the final approximate Paretooptimal set obtained by an MOEA. The function is as follows: where , is the mean of all , and is the number of nondominated solutions found so far. A value of zero for this metric indicates that all members of the Pareto front currently available are equidistantly spaced.
6.3. The Compared Algorithms
As Deb et al.’s suggestion in [13], some EMO procedures with variablewise recombination operators do not perform as well as those with vectorwise operators; besides, PCXNSGAII [13] and GDE3 [41] are recommended to handle linkagesbased multiobjective optimization problems. Zhang et al. [14] compared RMMEDA with GDE3, PCXNSGAII, and MIDEA [42]. They concluded that RMMEDA performed better than GDE3, PCXNSGAII, and MIDEA on some test instances with variable linkages, and MIDEA performed slightly badly among the four algorithms. Overall considering former works [13, 14], we choose RMMEDA, GDE3, and PCXNSGAII as the comparisons with MMEDA.
The third evolution version of generalized differential evolution (GDE3) is another updated version of original DE, which modifies earlier GDE version using a growing population and nondominated sorting with pruning of nondominated solutions to decrease the population size at each generation. The procedure is similar to NSGAII except that the simulator binary crossover is replaced with a differential evolution operator. The code of GDE3 used in comparison is written in MATLAB by the authors.
Parentcentric recombination (PCX) is a real parameter genetic operator in [43]. Deb et al. [13] introduced the PCX recombination operator in NSGAII and validated that PCXbased NSGAII performs better on some problem with variable linkages. The details of PCXNSGAII can be found in [13]. The code of PCXNSGAII is programmed by ourselves by referring to the code of SBXNSGAII and G3PCX from KanGAL (http://www.iitk.ac.in//kangal/).
We have discussed so much about RMMEDA in the former section in our paper. Here, we acknowledge the help of Professor Qingfu Zhang and Dr. Aimin Zhou for sharing the code of RMMEDA with us and their insightful comments.
6.4. Experimental Setup
In our experiments, the code is programmed in MATLAB. The source code of MMEDA can be obtained from the authors upon request. All the simulations run at a personal computer with PIV 3.0 G CPU and 2 G RAM. The experimental setting is as follows. Firstly, in GDE3, both CR and F in the differential operator are set to be 1 for F1~F10 and AF1~AF3, which have been investigated by Deb et al. [13] and Zhang et al. [14] on MOPs with variable linkages. Furthermore, by referring to the involved literatures [44, 45], CR and F are set to be 0.2 for ZDTs and DTLZ2, DTLZ4, and DTLZ6 except 0.5 for DTLZ1 and DTLZ3. Secondly, in PCXNSGAII, σ in PCX is set to be 0.4 for all the test instances which work well in studies in [13]. Thirdly, in RMMEDA, all the parameters’ setting is the same as the original paper; that is, in local PCA algorithm, is set to be 5, and extension rate is 0.25. Finally, in MMEDA, the size of tournament pool, , is 30, tournament scale is 2, and the size of active subpopulation, , is 40 for all test instances. Besides, the crossover probability of and a mutation probability of (where is the number of decision variables for realcoded GAs) for test instances. We choose and , which are similar to Deb et al.’s setting in [6].
Furthermore, the fourth column in Table 1 is the number of total evaluations for each test instance, respectively, which is followed by some significant past studies in this area [9, 14, 37, 46]. Indexes of the different algorithms are shown in Table 2. The values for different test instances are described in Table 3, which are similar to MOEA [31] and ODEMO [47]. Besides, we select 500 for two objective problems and 1000 for three objective problems with evenly distributed points in Paretooptimal front, and these points are denoted by .


6.5. Experimental Results of Multiobjective Problems without Variable Linkages
Deb et al. [13] and Zhang et al. [14] have investigated and validated that GDE3, PCXNSGAII, and RMMEDA performed better than other EMOAs on multiobjective problems with variable linkages. However, whether these algorithms could get the same good performance on multiobjective problems without variable linkages is still not investigated so far. For this end, we give the experimental comparison of MMEDA, RMMEDA, GDE3, and PCXNSGAII on famous ZDT and DTLZ problems in this section.
Figure 3 shows the Pareto fronts obtained from our algorithm in a random single run. These problems without variable linkages are rarely tested by estimation of distribution algorithms, and we can see that our method could get fairish results on famous ZDT and DTLZ test instances. Besides, seeing DTLZ2, DTLZ3, and DTLZ4 in Figure 3, we can gain that some extreme points of the Pareto front, as well as points located in segments of the Pareto front that are almost horizontal or vertical, are lost. This is the curse of dominance, which has been investigated by Deb et al. [31] and HernándezDíaz et al. [48].
Next, we investigate statistical results of the four algorithms in our paper in 30 independent runs on each test problem, which are in the form of box plots [49]. In a notched box plot, the notches represent a robust estimate of the uncertainty about the medians for boxtobox comparison. Symbol “+” denotes outlier (Figure 4).
The 30variable ZDT1 problem has a convex Paretooptimal front, while ZDT2 has a nonconvex Paretooptimal front, and ZDT3 provides difficulties by its discontinuities. Many MOEAs have achieved very good results on these problems in both goals of multiobjective optimization (convergence to the true Pareto front and uniform spread of solutions along the front). The results of the problems ZDT1, ZDT2, and ZDT3 (Figures 3 and 4) show that MMEDA achieves good results, which are comparable to the results of RMMEDA, GDE3, and PCXNSGAII; however, MMEDA is not the best in all metrics of the four algorithms. On the first three test problems we cannot see a meaningful difference in performance of the four algorithms. If it is exigent to find which is best, we can see that RMMEDA is slightly better than the other three algorithms in terms of spacing, while it performs a little poor in convergence metric. Besides, MMEDA and GDE3 are a little better in terms of convergence metric, while they achieve a little worse in spacing metric of these three problems.
ZDT4 is a hard optimization problem with many (21^{9}) local Pareto fronts that tend to mislead the optimization algorithm. In Figure 5, we can see that RMMEDA has difficulty in locating the global true Paretooptimal front. Besides, MMEDA, GDE3, and PCXNSGAII produce a very similar convergence and diversity measure. Furthermore, it seems that MMEDA demonstrates the best in terms of convergence and diversity by box plots in small scale.
With respect to the 10variable test problem ZDT6, there are two major difficulties. The first one is thin density of solutions towards the Pareto front and the second one is nonuniform spread of solutions along the front. The Paretooptimal solutions of DTLZ1 with (11^{5} – 1) local Pareto fronts lie on a threedimensional plane satisfying (Figure 3). In Figure 5, we can obtain that MMEDA and GDE3 achieve good convergence and diversity measures of ZDT6, followed by PCXNSGAII and RMMEDA, while, for DTLZ1, it seems that MMEDA gives both the best convergence and diversity measure in the four algorithms.
On the whole, we can obtain that MMEDA seems to be the best algorithm on these three problems. However, if we do not employ local searcher, EDAs based on probability global statistical information may perform weakly in approximating the Paretooptimal fronts of the three problems.
Problems of DTLZ2, DTLZ3, and DTLZ4 are threeobjective test instances with spherical Paretooptimal front (see Figure 3). Note that DTLZ3 has lots of local Pareto fronts, DTLZ4 emphasizes nonuniformity, and DTLZ6 has 2^{19} disconnected Paretooptimal regions in the search space. Figure 6 shows the performance metrics of the four algorithms on the four problems, respectively. We can find that MMEDA is the best in terms of both convergence and diversity on DTLZ3 and DTLZ4. Furthermore, for DTLZ3, MMEDA performs much better than the other there algorithms even though DTLZ3 has (3^{10} – 1) local Paretooptimal fronts. Gong et al. [9], Deb et al. [37], and Khare et al. [46] claimed that, for DTLZ3, NSGAII, SPEA2, and PESAII could not quite converge on to the true Paretooptimal fronts in 500 generations (50 000 function evaluations). In our paper, we have found that GDE3, PCXNSGAII, and RMMEDA also did badly in solving DTLZ3.
For convergence metric, it seems that GDE3 achieves the best measure in the four algorithms on DTLZ2 and DTLZ6. Besides, for diversity metric, MMEDA is the best on the four problems, which is the strongpoint of dominance. Since the dominance does not allow two solutions with a difference of in the th objective to be mutually nondominated to each other, it will be usually not possible to obtain the extreme corners of the Paretooptimal front. Although there is such shortcoming of dominance, we could still get better spacing measurement than that of RMMEDA, GDE3, and PCXNSGAII, which are based on crowding distance proposed by Deb et al. [6].
Overall considering the ten famous two and three objective problems without variable linkages, we can conclude that MMEDA produces a good convergence and diversity metrics with the exception of ZDT1, ZDT2, and ZDT3. GDE3 obtains comparative results for most of problems except DTLZ1 and DTLZ3. Besides, PCXNSGAII and RMMEDA are not better than MMEDA and GDE3 on most of the ten problems. However, the mechanism of RMMEDA is based on global statistical information, and strongpoints of RMMEDA are discovering linkages of variables and holding the spread of the final solutions. Besides, there are no variable linkages in the former ten problems. Therefore, it is not fair for RMMEDA. In [14], it has been validated that RMMEDA performs better than GDE3 and PCXNSGAII on some multiobjective problems with variable linkages. In the next subsection, we will give experimental results of our algorithms compared with the other three algorithms on twelve multiobjective problems with variable linkages.
6.6. Experimental Results of Multiobjective Problems with Variable Linkages
In this subsection, experimental results of twelve multiobjective problems with variable linkage are presented. The representation of experimental results is similar to that of Section 6.5; that is, 30 independent runs are performed on each test problem. The statistical results of our selected three metrics are shown by box plots too.
Figure 7 shows the Pareto fronts obtained by our algorithm on F1~F10 and AF1, AF2 in a random single run. We can see that MMEDA obtains good results of F1, F2, F3, F4, F5, F6, F9, AF1, and AF2. However, MMEDA performs a little poor on F7, F8, and F10, especially on F10, of which MMEDA cannot converge to the true Paretooptimal fronts. In [14], RMMEDA is also trapped in local Pareto front on F10. Furthermore, for GDE3 and PCXNSGAII, they converge to a small region of global Pareto fronts of F10, which is called “genetic drift.” Next, we will give experimental comparison with MMEDA, RMEDA, GDE3, and PCXNSGAII by box plots for further research.
Since F1 and F5 are variants of ZDT1 by introducing linear and nonlinear variable linkages into them, respectively, we put the statistical results of them in the same figure (Figure 8) for comparison, which are similar to F2 and F6. Note that the metric inverted generation distance should be first choice for comparison because inverted generation distance can measure both convergence and spread. If we ignore inverted generation distance, we cannot get rational and comprehensive comparison. Taking F6, for example, if we only concentrate on the metric convergence and spacing, it seems that PCXNSGAII is the best choice for this problem. Nevertheless, analyzing F6D in Figure 8, the result of F6 is in a small region of whole Pareto optimal front, which is called “genetic drift” by Goldberg and Segrest [50] and Fonseca and Fleming [51]. Zhang et al. got the same results of PCXNSGAII on F6 [14]. For this end, inverted generation distance is foundational one among the three metrics.
In terms of convergence metric of F1 and F2, MMEDA obtains best performance, while it performs a little poor in spacing metric. Furthermore, MMEDA and RMMEDA perform better than GDE3 and PCXNSGAII on the four problems in terms of spread and convergence. Although PCXNSGAII seems better in spacing metric on F5 and F6, it experiences difficulty in spread.
With respect to F3 and F7, since they are variants of ZDT6 by introducing linear and nonlinear variable linkages into them, respectively, there are three major difficulties of them. The first two difficulties are the same as ZDT6 as described above, while the third one is linkage relations between variable mappings. Therefore, F3 and F7 have more difficulty than F1, F2, F5, and F6. Figure 9 shows the performance measures on F3 and F7. We can obtain that RMMEDA and MMEDA achieve better performance than GDE3 and PCXNSGAII in terms of spread, convergence, and uniformity. This might be the reason why GDE3 and PCXNSGAII have no efficient mechanism for discovering variable relation, yet RMMEDA and MMEDA hold the property by building the probability distribution model based on manifold distribution of Pareto set in decision space, which have been described in detail in Section 3. Finally, it seems that RMMEDA is slightly better than MMEDA on the two problems.
F4 and F8 are variants of DTLZ2, and Figure 10 is the statistical results of them. We can obtain that MMEDA is slightly better than other three algorithms in our paper. Besides, GDE3 shows comparative results on F4, followed by RMMEDA. In [14], Zhang et al. admitted that GDE3 slightly outperforms better than RMMEDA on some problems and pointed out that the reason might be that RMMEDA does not directly use the local information of previous solutions in generating new solutions. Here, we give one possible answer, MMEDA, the hybridization of RMMEDA and an efficient local searcher. For F8, GDE3 and PCXNSGAII both converge to a curve and several points, which is “genetic drift” in multiobjective optimization. Figure 11 is the result of a single run of GDE3, and PCXNSGAII has similar results with GDE3. In multiobjective optimization, genetic drift means that finite population tends to converge to small regions of efficient set.
F9 is a variant of multimodal with Griewank function, while F10 is a variant of multimodal with Rastrigin function, and all of them are introduced into nonlinear mapping on variables. Figure 12 is the statistical results of them by the four algorithms in our study. We can obtain that MMEDA and RMMEDA get very similar performance of F9, and they get better measurement than GDE3 and PCXNSGAII. However, with respect to F10, it seems that MMEDA and PCXNSGAII are the best two algorithms in terms of the three metrics. However, it is not true to them. In Figure 7 (F10), we can see that all the solutions obtained by MMEDA converge to a small region near the global Paretooptimal front, which is similar to PCXNSGAII [14]. They are trapped in “genetic drift.” Besides, the final Pareto fronts obtained by RMMEDA and GDE3 on F10 are far from the Paretooptimal front, which can be validated by inverted generation distance (F10D in Figure 12) and convergence metric (F10Y in Figure 12). They are stagnated at local Pareto fronts. In a word, we have to admit that all the four algorithms cannot solve the problem efficiently.
By comparison, we can see that MMEDA is best in convergence and spread on AF1 and AF2, followed by RMMEDA and GDE3. For PCXNSGAII (Figure 13), it cannot converge to the true Paretooptimal front of the two problems. From these two problems, we can see that MMEDA, which hybrids with ITLS, is a competitive algorithm for some problems.
However, overall considering the experimental results of the twelve problems with variable linkages, RMMEDA and MMEDA are two more efficient and effective algorithms than traditional EAs for most of the problems. For EDAs, this new class of algorithms generalizes genetic algorithms by replacing the crossover and mutation operators by learning and sampling the probability distribution of the promising individuals of population, and the relationships between the variables involved in the problem domain are explicitly and effectively captured and exploited. Besides, the regularity that the distribution of Pareto set in the decision space is a piecewise continuous ()dimensional manifold is hybrid into MMEDA and RMMEDA. Therefore, we can conclude that MMEDA and RMMEDA should be better GDE3 and PCXNSGAII. Besides, from two problems (F3, F7) we can see that RMMEDA is better than MMEDA in all metrics. However, in terms of convergence, MMEDA is slightly better than RMMEDA on nine problems (F1, F2, F4, F5, F6, F8, F10, AF1, and AF2) and the effectiveness of our ITLS seems to play a very significant role for the above nine problems. But, for seven problems (F1~F3, F5~F7, and F9), RMMEDA outperforms MMEDA in terms of spacing metric.
As we know, convergence to the true Paretooptimal front and maintenance of a welldistributed set of nondominated solutions are two challenges for multiobjective optimization community. However, if solutions obtained by an algorithm cannot converge to the true Paretooptimal fronts, they are stagnated in local optimal fronts. It is useless to consider uniformity. Besides, if solutions obtained by an EMOA only converge to a small region of the whole Paretooptimal front, they have trouble in “genetic drift.” We cannot make a comprehensive view of the EMOA if we only take convergence and spacing metrics into account. Since the inverted generation can measure both convergence and spread of solutions obtained by an EMOA, it is used as remedy for the shortcomings of generation distance and spacing metrics. Next, we judge the statistical results by combined metrics.
Finally, with respect to the twentytwo multiobjective problems with and without variable linkages, we can conclude the following.(1)In terms of convergence and spread, MMEDA achieves the best measurement among the four algorithms on ZDT4, DTLZ1, DTLZ3, DTLZ4, F1, F2, F4, F5, F8, F9, AF1, and AF2 out of the twentytwo problems while RMMEDA performs best on F3, F6, and F7. GDE3 does best on ZDT1, ZDT3, ZDT6, DTLZ2, and DTLZ6. PCXNSGAII seems to obtain the best result of ZDT2.(2)In terms of spread and spacing, MMEDA performs the best performance on ZDT4, ZDT6, DTLZ1, DTLZ2, DTLZ3, DTLZ4, DTLZ6, F4, F8, AF1, and AF2, while RMMEDA is the best on ZDT1, ZDT2, ZDT3, F1, F3, F5, F6, F7, and F9. GDE3 achieves the best on F2.
Overall considering the three metrics, we can get that MMEDA turns out to be the best compromise among the four MOEAs considered in this study. However, MMEDA is possibly trapped in “genetic drift,” and we can get it from the results of F10 in Figure 3. Therefore, much work is needed to balance the exploration ability of global model and the exploitation ability of local searcher, or more efficient EDAs based on other machine learning techniques should be proposed.
In one word, depending on these empirical results and detailed analysis, we can draw that MMEDA is an efficient algorithm in solving multiobjective optimization problems with and without variable linkages.
7. Concluding Remarks
We have proposed a novel multiobjective algorithm based on the manifold distribution of Pareto set in the decision space, an efficient local searcher (ITLS), and dominance. Besides, two more difficult multiobjective problems by introducing nonlinear mapping into variable linkages are proposed. Following Deb et al. and Zhang et al.’s recent review of current evolutionary multiobjective optimization fields [13, 14] multiobjective problems with variable linkages have more difficulty for most current stateoftheart MOEAs. Consequently, in order to test our proposed algorithm, MMEDA, twentytwo multiobjective problems with and without variable linkages are employed, and RMMEDA, GDE3, and PCXNSGAII are used in comparison. The results show that our algorithms get competitive results in terms of convergence and diversity metrics.
However, there is much work to do in multiobjective optimization community. An issue that should be addressed in the future research is dominance. In our study, it seems that spacing metric of some twoobjective problems (such as ZDT1 and ZDT3) is not better than PCXNSGAII and GDE3, which are based on crowding distance in SBXNSGAII [6], and Deb et al. [31] have validated by experiments that MOEAs get better spacing metric than NSGAII based on crowding distance. There are two major reasons of this. The first is that we introduce a local searcher (ITLS) and the number of iterations will be less than the other three algorithms in our study under the same limit of total evaluations. The second is that some extreme points of the Pareto front, as well as points located in segments of the Pareto front that are almost horizontal or vertical, are lost. This is the curse of dominance. Some scholars have proposed Pareto adaptive dominance [48], which is an advanced version of dominance. But the loss of extreme solutions still exists. So much work is still needed to propose new version of dominance.
Another important topic is “genetic drift,” which has been studied by some researchers, such as Goldberg and Segrest [50], Fonseca and Fleming [51], and Srinivas and Deb [52]. At first, fitness sharing techniques are proposed to prevent genetic drift in multimodal function optimization and multiobjective optimization [52]. Later on, fitness assignment and crowding distance are proposed by Deb et al. [6] and Zitzler et al. [7] for diversity maintenance. Nowadays, some stateoftheart MOEAs have still trouble in “genetic drift” on some multiobjective problems with variable linkages [13, 14]. So how we can design new efficient and effective scheme to maintain diversity is an important issue of modern EMO community.
How to advance the local search ability of modern MOEAs is another worthwhile work. Gong et al. [9], Deb et al. [37], and Khare et al. [46] validated that, for DTLZ3, NSGAII, SPEA2, and PESAII, which are representative of stateoftheart MOEAs in multiobjective optimization community, could not quite converge on to the true Paretooptimal fronts in 500 generations (50 000 function evaluations). We have found that GDE3, PCXNSGAII, and RMMEDA also did badly in solving DTLZ3. DTLZ3 may be only one case of problems with many local Pareto fronts, and, possibly, there are many complicated problems like DTLZ3. Therefore, it requires designing more efficient and effective algorithms. Memetic algorithm may be one potential answer to the problem, and several multiobjective memetic algorithms have been proposed by some scholars [27, 29, 53], which have shown comparative performance than traditional EMOAs. A more efficient and effective hybrid algorithm should be addressed in the future research.
Conflict of Interests
The authors declare that there is no conflict of interests regarding the publication of this paper.
Acknowledgments
This work was supported by the National Natural Science Foundation of China (Grants nos. 61100173, 61100009, 61272283, 61201416, and 61304082) and the Foundation of Xi’an University of Technology (no. 116211306).
References
 E. Zitzler and L. Thiele, “Multiobjective evolutionary algorithms: a comparative case study and the strength Pareto approach,” IEEE Transactions on Evolutionary Computation, vol. 3, no. 4, pp. 257–271, 1999. View at: Publisher Site  Google Scholar
 J. D. Knowles and D. W. Corne, “Approximating the nondominated front using the pareto archived evolution strategy,” Evolutionary Computation, vol. 8, no. 2, pp. 149–172, 2000. View at: Publisher Site  Google Scholar
 D. W. Corne, J. D. Knowles, and M. J. Oates, “The Paretoenvelope based selection algorithm for multiobjective optimization,” in Parallel Problem Solving from Nature PPSN VI, vol. 1917 of Lecture Notes in Computer Science, pp. 869–878, Springer, New York, NY, USA, 2000. View at: Publisher Site  Google Scholar
 D. W. Corne, N. R. Jerram, J. D. Knowles, and M. J. Oates, “PESAII: regionbased selection in evolutionary multiobjective optimization,” in Proceedings of the Genetic and Evolutionary Computation Conference (GECCO '01), pp. 283–290, Morgan Kaufmann Publishers, San Francisco, Calif, USA, 2001. View at: Google Scholar
 C. A. Coello Coello and G. T. Pulido, “Multiobjective optimization using a microgenetic algorithm,” in Proceedings of the Genetic and Evolutionary Computation Conference (GECCO '01), pp. 274–282, Morgan Kaufmann, San Francisco, Calif, USA, 2001. View at: Google Scholar
 K. Deb, A. Pratap, S. Agarwal, and T. Meyarivan, “A fast and elitist multiobjective genetic algorithm: NSGAII,” IEEE Transactions on Evolutionary Computation, vol. 6, no. 2, pp. 182–197, 2002. View at: Publisher Site  Google Scholar
 E. Zitzler, M. Laumanns, and L. Thiele, “SPEA2: improving the strength Pareto evolutionary algorithm,” in Evolutionary Methods for Design, Optimization and Control with Applications to Industrial Problems, pp. 95–100, Athens, Greece, 2002. View at: Google Scholar
 C. A. Coello Coello, G. T. Pulido, and M. S. Lechuga, “Handling multiple objectives with particle swarm optimization,” IEEE Transactions on Evolutionary Computation, vol. 8, no. 3, pp. 256–279, 2004. View at: Publisher Site  Google Scholar
 M. Gong, L. Jiao, H. Du, and L. Bo, “Multiobjective immune algorithm with nondominated neighborbased selection,” Evolutionary Computation, vol. 16, no. 2, pp. 225–255, 2008. View at: Publisher Site  Google Scholar
 Y. T. Qi, X. L. Ma, F. Liu, L. C. Jiao, J. Y. Sun, and J. S. Wu, “MOEA/D with adaptive weight adjustment,” Evolutionary Computation, vol. 22, no. 2, pp. 231–264, 2014. View at: Google Scholar
 A. Zhou, B.Y. Qu, H. Li, S.Z. Zhao, P. N. Suganthan, and Q. Zhangd, “Multiobjective evolutionary algorithms: a survey of the state of the art,” Swarm and Evolutionary Computation, vol. 1, no. 1, pp. 32–49, 2011. View at: Publisher Site  Google Scholar
 C. A. Coello Coello, “Evolutionary multiobjective optimization: a historical view of the field,” IEEE Computational Intelligence Magazine, vol. 1, no. 1, pp. 28–36, 2006. View at: Publisher Site  Google Scholar
 K. Deb, A. Sinha, and S. Kukkonen, “Multiobjective test problems, linkages, and evolutionary methodologies,” in Proceeding of the 8th Annual Genetic and Evolutionary Computation Conference (GECCO '06), pp. 1141–1148, Washington, DC, USA, July 2006. View at: Google Scholar
 Q. F. Zhang, A. M. Zhou, and Y. C. Jin, “RMMEDA: a regularity model based multiobjective estimation of distribution algorithm,” IEEE Transactions on Evolutionary Computation, vol. 12, no. 1, pp. 41–63, 2008. View at: Publisher Site  Google Scholar
 N. Khan, Bayesian optimization algorithms for multiobjective and hierarchically difficult problems [M.S. thesis], University of Illinois at UrbanaChampaign, Urbana, Ill, USA, 2003.
 M. Pelikan, D. E. Goldberg, and E. CantPaz, “BOA: the Bayesian optimization algorithm,” in Proceedings of the Genetic and Evolutionary Computation Conference (GECCO '99), W. Banzhaf, Ed., Illigal Report No. 99003, pp. 525–532, Morgan Kaufmann, San Francisco, Calif, USA, 1999. View at: Google Scholar
 M. Pelikan, Hierarchical Bayesian Optimization Algorithm: Toward a New Generation of Evolutionary Algorithms, Springe