Abstract

The Pareto dominance-based evolutionary algorithms can effectively address multiobjective optimization problems (MOPs). However, when dealing with many-objective optimization problems with more than three objectives (MaOPs), the Pareto dominance relationships cannot effectively distinguish the nondominated solutions in high-dimensional spaces. With the increase of the number of objectives, the proportion of dominance-resistant solutions (DRSs) in the population rapidly increases, which leads to insufficient selection pressure. In this paper, to address the challenges on MaOPs, a knee point-driven many-objective evolutionary algorithm with adaptive switching mechanism (KPEA) is proposed. In KPEA, the knee points determined by an adaptive strategy are introduced for not only mating selection but also environmental selection, which increases the probability of generating excellent offspring. In addition, to remove dominance-resistant solutions (DRSs) in the population, an interquartile range method is adopted, which enhances the selection pressure. Moreover, a novel adaptive switching mechanism between angle-based selection and penalty for selecting solutions is proposed, which is aimed at achieving a balance between convergence and diversity. To validate the performance of KPEA, it is compared with five state-of-the-art many-objective evolutionary algorithms. All algorithms are evaluated on 20 benchmark problems, i.e., WFG1-9, MaF1, and MaF4-13 with 3, 5, 8, and 10 objectives. The experimental results demonstrate that KPEA outperforms the compared algorithms in terms of HV and IGD in most of the test instances.

1. Introduction

Multiobjective optimization problems (MOPs) are ubiquitous in the real-world, such as wireless sensor networks [1], engineering design [2], workflow applications [3], and robot path planning [4]. In MOPs, the objectives usually conflict with each other. Generally, an MOP with conflicting objectives can be defined as follows: where is an -dimensional decision vector in the decision space Ω and is the number of objective functions. MOPs with more than three-objectives (i.e., ) are also referred to as many-objective optimization problems (MaOPs) [5, 6].

Generally, many-objective evolutionary algorithms (MaOEAs) still adopt the Pareto dominance as the criterion for selecting solutions. However, in high-dimensional spaces, MaOEAs may face two significant challenges due to the curse of dimensionality. The first challenge is the phenomenon of dominance resistance. It means that the number of nondominated solutions in the population rapidly increases as the number of objectives increases, which leads to the incomparability of solutions [7, 8]. The resulting solutions that may survive for many generations in the population are called dominance resistance solutions (DRSs). The presence of DRSs will further expand the search space, which causes the convergence deterioration in the local search space. The second challenge is the contradiction between maintaining population convergence and diversity [9, 10].

To address these challenges, many novel MaOEAs have been proposed [11, 12]. These methods can be broadly classified into the following four categories.

The first category involves the redefining or relaxing the Pareto dominance relationship. Examples of improved dominance relationships include angle dominance [13], grid dominance [14], scalarization-based dominance relationship [15], RPS-dominance relationship [16], and reinforced dominance relationship [17]. NSGA-II+AD [13] proposes an angle dominance criterion that is insensitive to parameters. And the involved angle is composed of the solution and each objective axis. This dominance relationship cleverly expands the dominance region of a solution, which could increase the convergence pressure. In CDR-MOEA [18], a dominance relationship based on a convergence metric (CDR) is proposed, which combines the convergence metric and the adaptive parameter λ based on cosine similarity. This dominance relationship effectively balances the convergence and diversity of the nondominated solution set. MOEA/D-AED [19] uses adaptive Epsilon dominance to control the number of nondominated solutions, which effectively addresses problems with complex Pareto fronts (PFs).

The second type involves the decomposition-based algorithms, whose core idea is to solve MOP by decomposing a MOP into many single-objective subproblems. MOEA/D [20] is a typical decomposition-based algorithm that guides candidate solutions towards the Pareto front by defining a set of uniformly distributed weight vectors. However, this approach relies on the degree of match between the adopted reference vectors and the true Pareto front. To address the problems with irregular PFs, various adaptive decomposition-based methods have emerged. For example, AMOEA/D [21] adopts an adaptive addition-first-and-deletion-second strategy to dynamically adjust the weight vector. In addition, the reference points are widely used in most decomposition-based algorithms, such as NSGA-III [22], MOEA/DD [23], and SPEA/R [24]. SPEAR proposes a density estimation technique based on reference vectors to maintain population diversity. NSGA-III, as an extension of NSGA-II [25], features a combination of nondominated sorting with reference vectors. MOEA/DD, as a variant of MOEA/D, introduces the Pareto dominance and adaptively associates solutions with subregions. In summary, the decomposition-based method is suitable for solving MaOPs, but it is inevitable to preset the appropriate weight vectors and the scalar functions for guaranteeing their performance.

The third type is the indicator-based algorithm. Unlike the Pareto-dominance and decomposition-based algorithms, this method usually evaluates the quality of solutions using predefined performance indicators. Some well-known indicators are used as criteria in environmental selection, including inverted generational distance (IGD), hypervolume (HV), and generational distance (GD). The indicator-based method can provide more selection pressure towards the Pareto front. Some recently popular indicator-based algorithms include HypE [26], R2HCA-EMOA [27], and MaOEA/IGD [28]. HypE adopts the Lebesgue measure to calculate the fitness values of solutions. To reduce computational complexity, R2HCA-EMOA uses the R2 indicator to approximate the value of HV. However, this method cannot effectively balance convergence and diversity. MaOEA-IGD proposes a novel nondominance comparison strategy based on the IGD indicator. This strategy compares solutions to reference points, which improves the quality of population mating.

The final type is knee point-driven algorithms, which are aimed at improving the performance of MaOEAs by identifying knee points that contribute to a larger HV [29]. These algorithms employ strategies such as reflex angle [30], extended angle dominance [8], niching-based method [31], distance-based strategies [32], and offspring generation strategy [33] to identify knee points, which increase selection pressure and improve convergence. KnEA [34] identifies knee points based on the distance from the solutions to the hyperplane. And the number of knee points involved in this strategy is controlled by threshold parameter . However, the performance of KnEA is greatly influenced by the variation of parameter T. Pi-MOEA [35] uses the average ranking method to identify pivot solutions within the neighborhood, which is insensitive to parameters. k-NSGAII [36] applies a distance-based metric to identify convex and edge knee points. To improve the accuracy of the knee-detection method, k-NSGAII designs a specific parameter to control the width of the knee region. However, these strategies emphasize convergence too much but neglect the importance of diversity. In contrast, KnMAPIO [29] designs a novel environmental selection based on knee point-oriented dominance, which uses extreme and boundary points to select solutions from the critical layer. This selection strategy effectively balances the convergence and diversity.

In the past few years, many novel strategies for addressing MaOPs have been proposed. However, most MaOEAs still cannot meet the performance requirements. There are two main challenges.

First, without any prior knowledge, the diversity-first-convergence-second principle adopted by most Pareto dominance-based MaOEAs leads to the loss of population convergence.

Second, a strategy that overly emphasizes diversity may select many DRSs into the next-generation population, which leads to the population gradually moving away from the PF.

To address the above challenges, a knee point-driven many-objective evolutionary algorithm with adaptive switching mechanism (KPEA) is proposed in this paper, which is aimed at achieving a balance between convergence and diversity. Specifically, to increase the probability of generating excellent offspring, KPEA incorporates the knee points and the weighted distances into its mating strategy. During the environmental selection, an adaptive switching mechanism between angle-based selection and penalty is designed to balance convergence and diversity. The main contributions of this study can be summarized as follows: (1)Unlike the methods that indirectly identify DRS based on extreme solutions, to detect and eliminate DRSs in the population, an interquartile range method based on population convergence is used, which enhances the quality of population mating and ensures even scaling of each objective(2)Compared with selection mechanisms based on convergence-first or diversity-first, in the environmental selection process, an adaptive switching mechanism between angle-based selection and penalty can maintain a balance between convergence and diversity, instead of excessively emphasizing either of them

The remainder of this paper is organized as follows. The related works and the motivation of this study are presented in Section 2. A detailed description of the framework and core strategies of KPEA is presented in Section 3. The experimental setup is presented in Section 4. The experimental results and related analysis are presented in Section 5. Further investigation and discussion on KPEA are provided in Section 6. Finally, the conclusions of this paper and future work are given in Section 7.

In this section, the relevant research on dominance resistance solutions (DRSs) and angle-based selection strategies are reviewed, and the research motivation is elaborated briefly.

2.1. Impact of DRSs

In general, the Pareto-based algorithms regard the nondominated solutions as suitable candidates for the next generation. When some solutions are barely dominated but clearly inferior to others, then they will not be removed by the Pareto dominance-based strategies [7, 37, 38]. This phenomenon is called dominance resistance, and these solutions are called DRSs.

Regarding the specific definition of DRSs, various researchers have provided various descriptions. In literature [7], “DRSs are extremely inferior to others in at least one objective, and therefore, they are apart from Pareto optimums.” DRSs were rephrased in literature [38] as “These are points in the nondominated set that are extremely poor in at least one of the objectives, while being extremely good in some others.” Since DRSs are located on certain boundaries of the objective space, most may have the best density values, which causes these solutions to be always reserved for the next generation [39]. Furthermore, the proportion of nondominated solutions in a population grows exponentially as the number of objectives increases [40]. Meanwhile, the number of DRSs may also increase with the number of objectives, which results in more intense dominance resistance [5]. Even worse, DRSs can seriously inflate the nadir in specific axial directions, leading to an uneven scaling of each objective [8].

Figure 1 illustrates the situation of DRSs in the population in biobjective and triobjective spaces. The solid purple dots represent the nondominated solutions in the search space. In Figure 1, each axis shows the fitness value on each objective, and the nondominated solutions within the red dashed box are DRSs. With the increase of the number of objectives, the presence of DRSs will further expand the search space, which causes the convergence deterioration in local search space.

Some researchers have tried to identify the extreme solutions directly with specific methods. In many-objective particle swarm optimization (MaOPSO) [41], the solutions in the external archive that minimize the achievement scalarizing function (ASF) [42] are considered as the extreme solutions. In the multiphase balance of diversity and convergence in many-objective optimization (B-NSGA-III) [43], the extreme-LS operator is used to identify the extreme solutions in objective directions. Nevertheless, because MaOPSO and B-NSGA-III are all Pareto-based algorithms, their effectiveness is significantly impacted by DRSs. Different from directly determining the precise extreme solutions, some researchers are committed to deleting the DRSs in the current population to indirectly determine the extreme solutions.

Bhattacharjee et al. [38] propose a six-sigma-based method to eliminate DRSs. In this method, a solution is classified as DRSs when its value on the objective exceeds the average value plus six times the standard deviation of all nondominated solutions on that objective. However, the strategy still has some drawbacks. Xiang et al. [44] demonstrate experimentally that the method produces solutions with very poor convergence for some problems. Furthermore, Zhou et al. [45] propose a method to eliminate DRS by modifying the current candidate extreme solutions. Unfortunately, when there are several DRSs on an objective axis, this strategy cannot eliminate all DRSs.

In summary, the elimination of DRSs is beneficial to finding the exact extremal solutions. As Bhattacharjee et al. stated in literature [38], “the detection and elimination of DRSs is far from trivial, and further research is needed to address this issue.” In the paper, an interquartile range method (IQR) is introduced to eliminate certain nondominated solutions in the population whose convergence values are greater than the upper third quartile (Q3).

2.2. Angle-Based Selection Strategy

Basing on the embedding position of the angle selection criterion, the angle-based evolutionary algorithm can be classified into three categories. The first category involves applying the angle selection to environmental selection. The decomposition-based ranking and angle-based selection evolutionary algorithm (MOEA/D-SAS) [46] employs the angle information between solutions in the objective space to maintain population diversity. The vector angle-based evolutionary algorithm (VaEA) [47] uses the maximum-vector-angle-first strategy to select elite individuals. However, in certain cases, the selected solutions may have poor convergence. Thus, VaEA adopts the worse-elimination principle and conditionally uses alternative solutions to replace poorly converging ones. The scalar projection and perspective-based evolutionary algorithm (PAEA) [45] uses the minimum angle formed between a solution and a unit vector as a control threshold to maintain a balance between convergence and diversity. However, how to determine the threshold for the search direction in these strategies is usually an extremely complex problem.

The second category involves incorporating angle selection into the dominance relationship. Wang et al. [37] propose the generalized Pareto optimum (GPO) that uses a parameter , the extension angle on the th objective to redefine the dominance relationship. Yuan et al. [13] propose an angle dominance criterion that is insensitive to parameters. The involved angle is composed of the solution and each objective axis. This dominance relationship cleverly expands the dominance region of a solution, which could increase the convergence pressure.

The third category involves considering angle selection in both environmental selection and dominance relationships. He and Yen [48] propose a coordinated selection strategy to solve MaOP problems. This strategy uses ASF distance as the convergence metric and the angle formed between individuals as the diversity metric, which is applied to both environmental selection and mating selection. During the process of evolution, this strategy can balance the convergence and diversity of the whole population. Li et al. [49] use an archive population to update reference points and construct a distance scaling function based on the angle between individuals and reference points. This function is employed in mating selection, which improves the quality of offspring in the population.

Based on the above research, this paper presents a novel environmental selection strategy that regards the relationship between population size and the number of nondominated solutions as two criteria for selecting the solutions. The knee points determined by an adaptive strategy are introduced for not only mating selection but also environmental selection.

2.3. Motivation of This Paper

When dealing with many-objective optimization problems (MaOPs), dominance-based algorithms often face difficulties caused by DRSs and active diversity promotion [50, 51]. Decomposition-based algorithms can avoid the problem of insufficient selection pressure but require a predefined set of weight vectors. However, uniformly distributed weight vectors cannot guarantee well-distributed solutions. In contrast, angle-based selection strategies perform well on enhancing population diversity, especially when dealing with MaOPs with complex Pareto fronts. The studies in Section 2.2 indicate that many algorithms using angle-based ecological niche protection techniques seem effective, but there are still many unresolved problems. When dealing with nonlinear optimization problems, for the strategy proposed by the hyperplane-assisted evolutionary algorithm (hpaEA) [52] to protect prominent solutions, it may be invalid to maintain the diversity of the population.

The strategy of protecting prominent solutions proposed by the hyperplane-assisted evolutionary algorithm (hpaEA) [52] is not suitable for MaOPs with nonlinear Pareto fronts. On this type of problem, it is unable to maintain the diversity of the population for this strategy. The strategy proposed by VaEA overemphasizes diversity and lacks a reasonable design for angle thresholds, which results in difficultly achieving effective convergence in certain high-dimensional problems.

To address the above issues, this paper proposes a knee point-driven many-objective evolutionary algorithm with adaptive switching mechanism (KPEA). It incorporates the knee points and weighted distance into its mating strategy to increase the probability of generating excellent offspring. An interquartile range method is introduced to detect and eliminate DRSs in the population. During the environmental selection, an adaptive switching mechanism between angle-based selection and penalty is designed to balance convergence and diversity.

Up to now, most selection mechanisms focus on either convergence or diversity optimization. Generally, without any prior knowledge, a single strategy cannot offer a universal and acceptable performance on most MaOPs.

With this motivation, an adaptive switching mechanism between angle-based selection and penalty proposed in this paper could keep a balance between convergence and diversity, instead of excessively emphasizing either of them. The selection mechanism first preserves knee points identified by adaptive recognition and then selects individuals based on the maximum-vector-angle-first principle. The adaptive penalty mechanism selects the individuals with the best convergence one by one. Then, it adaptively adjusts the penalty set based on the distribution of solutions, which achieves comprehensive coverage of the Pareto front. As mentioned above, KPEA not only eliminates the influence of DRSs throughout the entire evolutionary process but also achieves a balance between diversity and convergence.

3. Proposed Algorithm

This section provides a detailed description of the proposed KPEA. Firstly, the main framework of the algorithm KPEA is given. Subsequently, the quartile detection method is regarded as a preprocessing tool to detect DRSs, which enhances the quality of population mating and ensures even scaling of each objective. In addition, in the mating selection, knee points and weighted distances are introduced as secondary selection criteria. Finally, an environment selection strategy is proposed with an adaptive switching mechanism, which can adaptively switch according to the relationship between population size and the number of nondominated solutions.

Input:P (population), N (population size), M (number of objectives), K (set of knee points), T (rate of knee points in population)
1 r ← 1, t ← 0 /Adaptive parameters for finding knee points/
2 K ←
3 PInitialize(N)
4 While termination criterion not fulfilled do
5     P’ ← Mating_selection(P, K, N)
6     QVariation(P’)
7     SPQ
8     P = Preprocessing(S) / A preprocessing operation to eliminate the DRSs. /
9     FNondominated_sort(P) /the preprocessed population is assigned to NF non-dominated fronts, Fi, 1 ≤ iNF/
10     [K, r, t] ← Finding_ Knee_ point(F, T, r, t)
11     PEnvironmental_selection(F, K, N)
12 End While
13 ReturnP
3.1. General Framework

A common Pareto-based MaOEA framework is employed in the proposed KPEA, as shown in Algorithm 1. First of all, an initial parent population of size is randomly generated. Then, after individuals are selected from using the binary tournament selection strategy, an offspring population is generated using mutation (lines 5 and 6). Three metrics, namely, dominance relation, knee point criterion, and weighted distance metric, are employed as the selection criteria for the binary tournament selection. If there are some checked DRSs in the combined population , a preprocessing mechanism based on IQR is used to detect and eliminate DRSs (line 8). Next, the nondominated sorting is performed on the preprocessed new combined population (line 9), and an adaptive strategy is introduced to identify the knee points of each nondominated front in the combined population (line 10). Finally, with an adaptive switching mechanism, individuals are selected from the population based on the relationship between population size and the number of nondominated solutions. As a vital contribution of this study, the niche protection operation different from VaEA [47] considers both convergence and diversity of each solution in the population.

In order to clearly illustrate the framework of KPEA, the specific flow of KPEA is shown in Figure 2. Specifically, KPEA consists of five evolution parts: mating selection, preprocessing, nondominated sorting, identification of knee points, and environment selection. During the environmental selection, an adaptive switching mechanism between angle-based selection and penalty is designed to balance convergence and diversity. In addition, the switching conditions and the specific evolutionary process of this novel adaptive switching mechanism are visualized.

3.2. Preprocessing Operation

Preprocessing is one of the most important steps in data mining, which is aimed at eliminating outliers [53]. In fact, during the optimization process of MaOPs, there are many DRSs (i.e., outliers) within the population. As discussed in Section 2, these DRSs have a negative impact on the quality of the final solutions. Thus, removing these DRSs is a crucial step to improve the performance of the algorithm.

In this paper, an interquartile range (IQR) method is introduced to detect and eliminate DRSs. IQR is a descriptive statistic for identifying outliers in any dataset. Given an ordered one-dimensional dataset , the lower quartile () and upper quartile () of the dataset are determined first, and then the IQR is calculated as follows. where and denote the upper and lower boundaries of the dataset, respectively. The value of parameter is set to 1.5, which will be comparatively analyzed in Section 4. In addition, and are defined as where indicates rounding up and represents the size of the set. The specific process of preprocessing aimed at the dataset , which is composed of convergence metrics of solutions, is visually illustrated in Figure 3.

Input:P (population), N (population size)
1 Compute the convergence metrics of solutions PCM
2 Sort CM;
3 The index position i is calculated, and then Q1 is calculated using Eq. (3).
4 The index position j is calculated, and then Q3 is calculated using Eq. (3).
5 Ifj > Nthen
6      and Q1 = CM(i)
7     j = N and Q3 = CM(j)
8 End
9 Compute the UB of the ordered set CM
10 Eliminate the solutions whose convergence metrics are greater than UB → P'
11 Calculate the disparity indicators Is of P and P': /Detect if there are DRSs in P/
12 Ifγ > 1 then
13 P = P'
14 End

First, the convergence metrics of all solutions are calculated and sorted in ascending order. According to the quartile method, the index positions and locate at 1/4 and 3/4 of the sorted convergence metrics, respectively. In Figure 3, and , where indicates rounding up. and are calculated according to Eq. (3). In Figure 3, and . Then, the lower and upper boundaries are set as are calculated according to Eq. (2). In this paper, a solution is considered a DRS when its convergence metric is greater than the upper boundary. Finally, the blue point with a convergence metric of 92 will be regarded as DRSs by the preprocessing mechanism and will be eliminated from the population.

Algorithm 2 presents the specific process of the preprocessing operation. In the first step, the convergence metric for each solution is calculated and stored in the set (line 1). Then, are sorted in ascending order (line 2). Subsequently, the corresponding index position of in is calculated according to the quartile method, and then the value of in the ordered is calculated using Eq. (3) (line 3). As the same as the solving procedure of , is calculated (line 4). To avoid the situation that the number of candidate solutions after preprocessing is less than the population size , and may be adjusted. If , the upper and lower quartiles of the are adjusted accordingly (lines 5-7). After determining the upper and lower quartiles, is calculated using Eq. (2) (line 9). Finally, any solutions with a convergence metric larger than are eliminated, leading to a new set of solutions (line 10). is an auxiliary set for detecting whether there are DRSs in . The disparity indicator is employed, which calculates the diagonal length of the hypercube formed by the extreme objective values in population (line 11). The disparity indicator is defined as where represents the number of objectives and is a candidate solution that belongs to the population. After calculating the disparity indicator and of and , the difference ratio between two sets is calculated as follows:

Obviously, if is smaller than or equal to 1, it implies that there are no DRSs in . On the contrary, if is greater than 1, it indicates that one or more DRSs exist in , and will be updated to with no DRSs (lines 12 and 13).

3.3. Adaptive Strategy of Identifying Knee Points

In KPEA, the knee points play a crucial role. According to literature [34], knee point-based selection has more advantages than using HV-based selection.

The core idea of determining the knee point is demonstrated in Figure 4. In biobjective minimization problems, the extreme solutions with the maximum value on the respective objective axes and need to be first determined. Then, the extreme line is defined as

Finally, the distance from each solution to is calculated as

Within the neighborhood range, the solution with the maximum distance to will be identified as the knee point. In many-objective optimization problems, refers to the hyperplane [54].

As shown in Figure 4, solutions , , and are located within the neighborhood enclosed by dashed lines. Among these, solution has the maximum distance to . Therefore, is considered as the knee point within the neighborhood. When there is only one solution within the neighborhood (e.g., solution ), that solution is also considered as the knee point.

The results of identifying knee points will be significantly affected by the size of the neighborhood of the solutions. Solutions , , and are identified as knee points based on the neighborhood size defined in Figure 4. However, when all solutions are included in the same neighborhood, only solution is identified as the knee point. To address this issue, a strategy is introduced to adjust the neighborhood size of solutions.

Assuming that at the th generation, there are -nondominated fronts within the population. Each nondominated front is composed of a set of nondominated solutions . The neighborhood of a solution is defined by a hypercube of dimension , where and denotes the number of objectives. Specifically, the neighborhood size of the th objective is calculated as follows: where and represent the minimum and maximum values of the nondominated set on the th objective at the th generation, respectively. The parameter denotes the ratio of the size of the neighborhood to the span of th objective in nondominated front at the th generation, which is defined as where represents the ratio of the size of the neighborhood to the span of th objective in at the ()th generation and indicates the ratio of knee points to the number of nondominated solutions in at the th generation. The threshold is used to control the proportion of knee points in .

The main steps for identifying the knee points are presented in Algorithm 3. The knee points are identified starting from the first nondominated front (lines 2). First, the extreme points in are identified, and the hyperplane is calculated (lines 3 and 4). Next, the size of the neighborhood for the solutions in is calculated (lines 5-8). The distances from each solution to L are calculated and sorted in descending order (lines 9&10). Then, the knee points within each neighborhood are identified and added into the knee point set (lines 12-15). Finally, the ratio between the number of knee points and the number of nondominated solutions in is updated (line 17).

Input:F (sorted population), T (rate of knee points in population), r, t (adaptive parameters)
1 K ← ∅ / knee points /
2 For allFiFdo
3   EFind_extreme_solution(Fi)
4   LCalculate_extreme_hyperplane(E)
5   Update r by Eq. (9)
6   fmax ← Maximum value of each objective in Fi
7   fmin ← Minimum value of each objective in Fi
8   Calculate V by Eq. (8)
9   Calculate the distance from each solution in Fi to L by Eq. (7)
10   Sort Fi in a descending order according to the distances
11   SizeFi
12   For allpFido
13      
14      KK ⋃ {p}
15      FiFiNB
16   End For
17   
18 End for
19 ReturnK, r and t
Input:P (population), K (set of knee points), N (population size)
1 Q ← ∅
2 Whiledo
3   Randomly choose x and y from P
4   Ifxythen
5     QQ ∪ {x}
6   Else Ifyxthen
7     QQ ∪ {y}
8   Else
9     IfxK and yKthen
10       QQ ∪ {x}
11     Else IfyK and xKthen
12       QQ ∪ {y}
13     Else
14       IfDW(x) > DW(y) then
15          QQ ∪ {x}
16       Else IfDW(x) < DW(y) then
17          QQ ∪ {y}
18       Else
19          Ifrand() < 0.5 then
20            QQ ∪ {x}
21          Else
22            QQ ∪ {y}
23          End If
24       End If
25     End If
26   End If
27 End While
28 ReturnQ
3.4. Binary Tournament Mating Selection

In KPEA, a binary tournament selection based on three metrics, including the Pareto dominance, knee point criterion, and weighted distance, is designed, as shown in Algorithm 4.

Two individuals, and , are randomly selected from the parent population . If is dominated by , then will be preferred (lines 4-7). If and are nondominated, the individual belonging to the knee point set will be selected (lines 9-12). If both or none of and belongs to the knee point set , the weighted distances will be used to compare and . The weighted distances of and are calculated as follows: where is the th nearest neighbor of in the population, is the weight of , is the Euclidean distance between and , and is the rank of distance among all distances , .

Then, the individual with a large value of is selected (lines 14-17). If , either is selected (lines 19-23). This procedure is repeated until individuals are selected.

3.5. Environmental Selection

In this paper, a novel environmental selection strategy is proposed, which is aimed at enhancing selection pressure while balancing convergence and diversity.

The details of the novel environmental selection strategy are provided in Algorithm 5. Firstly, the sorted population will be normalized (line 1). The objective vector of a solution in the sorted population can be normalized to using the following formula [55, 56]: where denotes the ideal point, that is, , ; represents the worst point, that is, , .

When the number of solutions in exceeds the population size , the solutions based on the Pareto dominance relationship are incomparable. In this situation, the convergence and diversity of KPEA are maintained alternately by the penalty mechanism (introduced in Section 3.5.1), which also decreases the sensitivity to the distribution threshold (lines 2 and 3).

In contrast, when the number of nondominated solutions in is less than the population size, knee points are first preserved by the selection mechanism (introduced in Section 3.5.2), which increases selection pressure and improves convergence. Then, the remaining solutions are selected with association and niching functions based on the principle of maximum-vector-angle-first, which maintains diversity (lines 4 and 5). The detailed explanations of these methods will be provided in the following sections.

Input:F (sorted population), K (set of knee points), N (population size);
Output: P (next population)
1 Normalization (F)
2 Ifthen
3   PPenalty(F1, N); / Solutions are selected one by one using penalty mechanism. /
4 Else
5   PSelection(F, K, N); / The knee points are first preserved, and then each solution is selected based on the principle of maximum-vector-angle-first. /
6 End
7 ReturnP
Input:F (sorted population), N (population size);
Output: P (next population)
1 whiledo
2   Compute the angles between piand {p1, p2, …, pi-1, pi+1, …, }
3   SDM = {DM1, DM2, …, }
4   Compute the distance CMi from pi to the idle point z;
5   SCM = {CM1, CM2, …, }
6   
7   Select pi with min(SCM)
8   Move pi into P
9   Identify the neighbors Ni of pi, where μij < α and μij ∈ {μi1, μi2, …, }
10   Move Ni into the penalty set E
11   IfSize(E) > SNthen
12     Identify the neighbors Nj of P, where the DMjof Nj is smaller compared to the other neighbors of P, DMj ∈ {DM1, DM2, …, }
13     Penalize Nj into the penalty set E
14     Move the remaining solutions Ntinto the solution set F, where the DMtof Nt is larger compared to the other neighbors of P, DMj ∈ {DM1, DM2, …, }
15   End
16 End
17 ReturnP
3.5.1. Penalty

When the number of solutions in is greater than the population size , the function is triggered in KPEA, as shown in Algorithm 6. The excellent solutions for the next generation are selected based on the penalty principle, which considers both the diversity and convergence of the population. First, the angles between and is computed and stored in in ascending order (lines 2-3). The distance from to the idle point is computed and stored in (lines 4-5). Then, the solution with the minimum value in is selected from the candidate solution set and put into the next-generation population (lines 7 and 8). Once a solution is selected, its neighbors are identified, where and . These neighbors are put into the penalty set (lines 9 and 10). The size of is determined by the size of the initial candidate set and the desired population size . If the number of solutions in exceeds its capacity limit, the penalty set must be updated based on the similarity between the penalized solutions and the selected solution set . Specifically, the closer the solution in is to the candidate solution, the smaller the angle value (stored in ) between and the candidate solution. Therefore, among the neighbors of the candidate solutions, the top neighbors with smaller angle values are kept in the penalty set (lines 11-13). Then, the remaining well-distributed solutions with larger angle values (stored in ) will be moved from to the candidate set (line 14).

The process of the penalty mechanism in a 2D objective space is illustrated in Figure 5. In this example, there are 9 nondominated solutions in , and the population size is 5. Firstly, solution with is selected. And then, as the angles between the solutions ( and ) and are smaller than , and will be penalized, as shown in Figure 5(b). Next, as the same as the method in Figure 5(b), solution with is selected. As the number of solutions in the penalty set does not exceed the limit, the neighbor of can be directly punished, as shown in Figure 5(c). Subsequently, solution with is selected. At present, solutions , , and are contained in the population . Solutions , , and are included in the penalty set. Because the penalty set can only accommodate four solutions, the neighbors and cannot be directly penalized by the mechanism. Therefore, it is necessary to update the penalty set by removing solution with . After the update, the penalty set is composed of solutions , , , and . The selected solution may have no neighbors, which means there is no angle between the candidate solutions and the selected solutions smaller than . In this case, no solution is penalized. Finally, based on the above steps, solutions and are selected, as shown in Figures 5(e) and 5(f).

3.5.2. Selection

Once the number of nondominated solutions in is less than the population size, the selection operation will be triggered, as given in Algorithm 7. The knee points of each nondominated front are selected first and added into , which can increase selection pressure (line 2). Then, the remaining nondominated solutions from are selected and added into . If the number of solutions in exceeds the population size , then is considered as the critical front. knee points with the minimum distance to the hyperplane are removed (lines 4 and 5). Otherwise, the association function that constructs the search direction based on and the niching function that is aimed at ensuring the diversity of the population are used to select solutions one by one from (lines 6–8). The association and niching functions will be described in detail in the following sections.

Input:F (sorted population), K (set of knee points), N (population size);
Output: P (next population)
1 P = , i = 1
2 PP ∪ (KFl) / Select the knee point and the solutions in F1Fl into P/
3 PF1 ∪… ∪ Fl-1
4 Ifthen
5   Delete solutions from KFl, which have the minimum distances to the hyperplane
6 Else Ifthen
7   Associate each member of Fl with a solution in P: Association (P, Fl)
8   Choose solutions one by one from Fl to construct final P: P = Niching (P, Fl, )
10 End If
11 ReturnP
Input: P (population), Fl (critical front)
1 For each xjFl
2   θ (xj) = ∞
3   γ (xj) = -1
4   For each ckP
5     Calculate angle μ between xj and ck by Eq. (15)
6     Ifμ < θ (xj)
7        θ (xj) = μ
8        γ (xj) = k
9     End If
10   End For
11 End For
3.5.3. Association

The angle-based association process is described in this section. After selecting the knee points into , if is not full, solutions will be selected from based on the vector angles. Before describing the association operation, some relevant formulas are first presented in this section.

The norm could calculate the vector angle between two solutions in the normalized objective space. The objective vector in the sorted population has been normalized to , as introduced at the beginning of Section 3.5. The norm of solution in the sorted population is defined as

The vector angle between solutions and is defined as where is to calculate the inner product between and . And is defined as

The main steps of the association process are provided in Algorithm 8. For each remaining member in , for example, , the minimum vector angle (denoted as ) between and the solutions in is calculated. The index of the solution in that has the minimum vector angle to is denoted as .

Firstly, for each , its and are initialized as ∞ and -1, respectively (lines 1-3). Then, for each , the angle μ between and is calculated (lines 4 and 5). If and are updated accordingly (lines 6–8). The association process in a 2D objective space is illustrated in Figure 6. In this example, there are 4 nondominated solutions, and one knee point has been selected and put into the next-generation population . Based on the vector angle association, both and are associated with , , and are associated with , and is associated with . Taking as an example, and are and 2, respectively.

Input: P (population), Fl (critical front), (solutions to be chosen)
Output:P (next population)
1 T =  /   returns the cardinality of Fl/
2 flag(xj) = false, xjFl, j = 1, 2, ..., T
3 Fork = 1 to ()/ Find the index with the maximum vector angle/
4   
5   P = Maximum-Vector-Angle-First (P, ρ, T)
6 End for
7 ReturnP
Input:P (population), ρ (solution index); T (number of solutions)
Output:P (next population)
1 Ifρ == null,
2   ReturnP
3 End If
4 P = P ∪ {ρ} /ρ is orderly added/
5 flag(ρ) = true
6 Forj = 1 to T
7   Ifflag(xj) == false /xj is a member of Fl/
8     Calculate angle between xj and ρ by Eq. (15)
9     Ifangle < θ (xj) /xj is associated with ρ/
10        θ (xj) = angle /Update the vector angle from xj to P/
11         /Update the corresponding index./
12     End if
13   End if
14 End for
15 ReturnP
3.5.4. Niche Preservation

A niche-based diversity maintenance strategy is described in Algorithm 9, which is based on the principle of maximum-vector-angle-first.

First, each solution in has a flag value, which indicates whether the solution has been added into . These flag values should be initialized to false at the beginning of the process (lines 1 and 2). Subsequently, for each , the index of the solution in that has the maximum vector angle to is calculated, which is denoted by (lines 3 and 4).

Finally, based on the principle of maximum-vector-angle-first, the best solution is selected and added to , as specified in Algorithm 10. When is null, it means that all solutions in have been added into . Otherwise, is added into and the corresponding flag is updated to true (lines 4 and 5). When adding to the population , the vector angles between the remaining solutions in and the solutions in must be adjusted accordingly. This adjustment can be achieved by computing the vector angle between and each solution in whose flag value is false (lines 7-11). The specific updating process has been described in Section 3.5.3. As shown in Figure 6, compared to the other remaining solutions, the angle between the solution in and in is the maximum. Based on the principle of maximum-vector-angle-first, will be added into . Obviously, compared to solutions such as , , and , adding is more likely to improve the distribution of the population. In addition, the potential solutions may be obtained along the direction of . Assuming that two more solutions need to be added into , and will be chosen.

In contrast, can never be considered because the vector angle between and the associated solution is zero. This means that with the same direction as has good convergence in .

3.6. Time Complexity Analysis

During each evolutionary generation, KPEA performs five main operations: mating selection, preprocessing, nondominated sorting, identification of the knee points, and environmental selection. The preprocessing operation takes a runtime of . Nondominated sorting operation takes a runtime of . The identification of the knee points requires at most a runtime of . Specifically, it takes at most a runtime of to compute the distance between each nondominated solution and the hyperplane. And it takes a runtime of to identify the knee points in the nondominated solutions set.

During the environment selection, the time complexity of KPEA is mainly determined by the adaptive switching strategy, as described in Algorithm 5. When the number of nondominated solutions in exceeds the population size , the function penalty operation is triggered. First, the computation of cosine similarity requires a runtime of . Next, it takes a runtime of to identify the elite solutions. When the number of nondominated solutions in is less than , the selection operation will be triggered. First, it takes a runtime of to identify the knee points into the next generation of the population. Next, the association and niching functions require at most a runtime of .

Therefore, the environment selection stage requires at most a runtime of . Overall, in each generation, the worst time complexity of KPEA is .

4. Experimental Setup

The experiment to validate the performance of the proposed KPEA on MaOPs is set up in this section. First, in Section 4.1, five representative algorithms are briefly introduced, namely, k-NSGAII [36], KnEA [34], Pi-MOEA [35], hpaEA [52], and VaEA [47]. Then, the specific characteristics of the test problems in the MaF and WFG benchmark suites are listed in Section 4.2. Finally, performance metrics and parameter settings are described in Section 4.3 and 4.4, respectively.

4.1. Comparison Algorithms

To fully validate the effectiveness of KPEA, five state-of-the-art algorithms are selected for performance comparison in this paper. The core ideas of these five algorithms are presented as follows. (i)k-NSGAII [36]. The algorithm is based on NSGA-II and introduces a knee point identification strategy, which uses a user-defined parameter to control the width of the neighborhood of knee points. This strategy can improve the coverage of the population on the Pareto front(ii)KnEA [34]. The algorithm adopts an adaptive strategy to identify knee points for each nondominated front and uses a custom threshold parameter to control the number of knee points recognized. This strategy not only enhances selection pressure on the Pareto front but also promotes population diversity(iii)Pi-MOEA [35]. The algorithm uses an average ranking method to identify pivot solutions within the adaptive neighborhood, which is aimed at achieving a balance between convergence and diversity(iv)VaEA [47]. The selection of solutions is based on the principles of maximum-vector-angle-first and worse-elimination without requiring additional weight vectors and parameters(v)hpaEA [52]. With the assistance of hyperplanes, prominent solutions are selected first during both mating selection and environmental selection, which increases the selection pressure

4.2. Description of Benchmark Problems

In this paper, 20 scalable test problems from two well-known benchmark test suites, WFG [57] and MaF [58], are used in this study to verify the performance of KPEA. The characteristics of the test problems are detailed in Table 1.

4.3. Performance Indicators

Two widely used performance metrics, namely, hypervolume (HV) [59] and inverse generational distance (IGD) [60, 61], are considered as they can simultaneously evaluate the convergence and diversity of solution sets. The IGD measures the average distance between the nondominated solution set and the nondominated solution set uniformly distributed along the true Pareto front (PF) [26]. IGD is defined as follows: where refers to the minimum distance between the solution in and its closest solution in . Therefore, the smaller the value of IGD, the closer the solution set is to the true PF.

refers to the volume of a hypercube in the objective space that is enclosed by a reference point and a set of nondominated solutions , which is a common indicator for measuring the performance of the algorithm. It can be defined as follows: where is the hypercube enclosed by and the individual in . The quality of the solution set can be judged by the value.

In order to obtain reliable statistical conclusions, the Wilcoxon rank-sum test [62] with a significance level of 0.05 is conducted to evaluate whether the proposed KPEA has a significant advantage over other comparative algorithms. The symbols “+,” “-,” and “=” indicate whether the results obtained by the compared algorithms are significantly better, similar, and worse than those obtained by KPEA, respectively. Moreover, the Friedman test [63] is employed in this study to obtain the average rankings of the algorithms on each test instance, which helps the overall performance of each algorithm.

4.4. Parameter Settings

The several key parameters adopted in this experiment are summarized as follows: (i)Variation Operators. In this paper, the simulated binary crossover (SBX) [64] and polynomial mutation [65] are adopted by all the MaOEAs used for comparison to generate offspring solutions. The crossover probability is set to 1, and the mutation probability is set to , where is the number of decision variables. Following the recommendations in literature [23], the distribution indexes and are set to 20 and 30, respectively. The other parameter settings for the algorithms are consistent with the settings described in their respective literature(ii)Population Size. This paper considers 3, 5, 8, and 10 objectives. The population size of six algorithms for different numbers of objectives is summarized in Table 2(iii)The Number of Runs and Termination Criterion. For each test instance, each algorithm runs 20 times independently. The termination criterion for each run is determined by the predefined maximum number of generations (maxGen). For WFG1 and WFG2, maxGen is set to 1000. For WFG3-WFG9, maxGen is set to 250. For the MaF test suite, maxGen is set to 250. When the number of generations reaches maxGen, all algorithms are terminated

5. Experimental Results

In this section, according to the experimental design described in Section 4, KPEA is compared with five state-of-the-art algorithms on WFG and MaF. The statistical results of IGD and HV metrics obtained by the six algorithms on the test suite are listed in Tables 36. The best results for each problem instance are annotated with a dark gray background. Then, in Section 5.3, the sensitivity of KPEA to parameter is discussed. Additionally, the performance comparison of KPEA and its different variants is analyzed in Section 5.4. Finally, in Section 5.5, the overall performance of six algorithms is analyzed.

5.1. Analysis of WFG Test Problems

The mean and standard deviation values of IGD and HV results obtained by five compared algorithms and KPEA on the WFG test suite are listed in Tables 3 and 4. The performance is challenged by the multiple complex features of WFG problems, which include scalability, deception, and separability.

As shown in Table 3, KPEA significantly outperforms other algorithms. Specifically, among 36 instances, KPEA outperforms all compared algorithms on 18 instances. The performance of KPEA is significantly better than the other five algorithms on the WFG1, WFG2, and WFG8 problems with scaling and concave-convex Pareto fronts. It can be attributed to the novel adaptive switching mechanism of KPEA, which balances diversity and convergence. In terms of the HV indicator, KnEA performs better on the WFG5 and WFG9 problems with deceptive and concave Pareto fronts. This is due to the fact that KnEA sets a small threshold for the adaptive strategy to identify knee points, which helps KnEA to achieve good performance on scalable problems. Pi-MOEA performs better than KPEA when solving the WFG problems with three-objective. Furthermore, the performance of VaEA is considerably worse than that of KPEA. The reason is that VaEA emphasizes diversity too much but the importance of convergence is neglected.

In Table 4, in terms of IGD values, KPEA and VaEA achieve better results than other algorithms on most of the WFG test problems. Out of the 36 test instances, KPEA achieves the best IGD result on 20 instances, while Pi-MOEA obtains the best result on 8 instances. The reason why KPEA performs well on some WFG test problems may be that it combines the knee points with the maximum-vector-angle-first principle to guide the evolution of the population. Therefore, unlike KnEA and k-NSGAII that lack a diversity maintenance mechanism, it is not easy for KPEA to trap in the local optima on problems with biased Pareto fronts such as WFG1, WFG8, and WFG9. Pi-MOEA uses a threshold defined based on the average ranking to control the neighborhood range of solutions, which helps the algorithm maintain a balance between convergence and diversity. It is important to note that on the WFG test suite, the performance of the k-NSGAII and hpaEA is relatively the worst in both HV and IGD metrics. For k-NSGAII, the effectiveness of the knee point-driven algorithms is heavily influenced by the approach used to detect the knee points, which is a crucial factor. However, the approach proposed by k-NSGAII only focuses on a small portion of the PF region when identifying the knee points, which results in poor diversity during the search process of k-NSGAII. For hpaEA, this is mainly due to its proposed prominent solution protection strategy, which is unable to maintain a good diversity of solutions on the linear Pareto front of the WFG test suite.

Taking WFG1 and WFG9 with 10 objectives as examples, the final solutions obtained by six algorithms in a single run are shown in Figures 7 and 8, respectively. For the WFG1 problem with 10 objectives, each algorithm encounters some difficulties in solving this problem, as shown in Figure 7. In terms of distribution and convergence, both hpaEA and VaEA perform poorly. Furthermore, the uniformity of the Pareto front obtained by k-NSGAII and KnEA is worse than that of KPEA. Only Pi-MOEA and KPEA show relatively good performance, because their obtained solutions are closer to the true Pareto front. From Figure 8, the solutions obtained by KPEA and KnEA are relatively better in terms of both convergence and distribution. While the solutions obtained by k-NSGAII, Pi-MOEA, and hpaEA have all been trapped in local optima on the Pareto front. The solutions obtained by VaEA are the worst in terms of both convergence and diversity.

5.2. Analysis of MaF Test Problems

Although WFG problems are widely used to evaluate the performance of many-objective optimization algorithms, their benchmark functions are too regular and monotonous. Furthermore, WFG problems lack some validation for the complex and uncertain real-world problems. In contrast, the MaF problems that focus on evaluating diversity and convergence have higher complexity and practical significance. In this study, tests are conducted on the MaF problems to further validate the performance of KPEA.

The statistical results of the HV values obtained by six algorithms on the MaF test suite are presented in Table 5. KPEA outperforms the other five algorithms on 27 instances out of the 44 test instances. Furthermore, KPEA also demonstrates strong competitiveness on MaF4 with a convex Pareto front, MaF8, MaF11 with a disconnected Pareto front, and MaF13 with irregular mixed Pareto fronts.

The IGD values obtained by the considered algorithms on the MaF test suite are presented in Table 6. Out of the 44 test instances, KPEA achieves the best IGD value on 26 instances. Additionally, Pi-MOEA also performs well, which obtains the best IGD value on 6 test instances. Conversely, the performance of k-NSGAII and KnEA is not as good as other algorithms.

To illustrate the convergence and distribution of solutions more intuitively, the final population of six algorithms on several representative problems is plotted in Figures 912. KPEA, Pi-MOEA, hpaEA, and VaEA maintain a uniform population distribution on MaF1 with an irregular PF shape, as shown in Figure 9. On the contrary, the distribution of solutions obtained by k-NSGAII and KnEA is relatively poor. The distribution of solutions obtained by different algorithms on the MaF5 with three objectives is shown in Figure 10. The solutions with the uniform distribution obtained by Pi-MOEA and KPEA are closest to the true Pareto front of MaF5. Compared to k-NSGAII and hpaEA, the solutions of KnEA and VaEA seem to have good convergence. However, both KnEA and VaEA struggle to maintain population diversity on MaF5. From Figure 11, the solutions obtained by KnEA, VAEA, and KPEA are closest to the true Pareto front when the number of objectives is 10. Moreover, as shown in Figures 10 and 11. unlike other algorithms, the performance of KPEA becomes more prominent as the number of objectives increases. MaF8 is a problem of multipoint distance minimization [43]. From Figure 12, KPEA demonstrates the best convergence and diversity performance on MaF8. However, the distribution of k-NSGAII, Pi-MOEA, and KnEA is quite unsatisfactory. The solutions obtained by hpaEA and VaEA cannot converge to the Pareto front on MaF8. This also proves the effectiveness of the proposed KPEA.

5.3. Sensitivity of Parameter

To investigate the sensitivity of parameter r of preprocessing in KPEA, the performance with varying from 0 to 5 on 3, 5, and 10 objective test problems is evaluated. Seven representative problems in Table 7 are selected for testing and comparison. The average IGD values in 30 iterations for each problem are calculated.

From Figure 13, on MaF2 and WFG9 problems with concave and scaled Pareto fronts, the performance of KPEA is insensitive to . On MaF1 with degenerate Pareto front, the analysis needs to be conducted based on the number of objectives. As shown in Figures 13(a) and 13(b), when the number of objectives is less than 10, the IGD value of KPEA slightly fluctuates as the value of increases. However, when the number of objectives reaches 10, the variation of has a significant impact on the performance of KPEA. Firstly, as increases from 0 to 1.5, the IGD value consistently decreases. Then, as continues to increase, the IGD value gradually increases. Moreover, on WFG1 and MaF9, the IGD value also slightly fluctuates as the value increases. However, no matter how the value fluctuates, the IGD values of KPEA are always the smallest at . On WFG2 and MaF6, the IGD values rapidly decrease until and then tends to stabilize with the increase of .

Therefore, is the best choice to improve the performance of KPEA. The parameter setting of can significantly enhance the applicability of the algorithm.

5.4. Performance Comparison of KPEA and Its Different Variants

The performance of KPEA is compared with its three variants in this section. The experimental results are used to validate the effectiveness of penalty, the effectiveness of preprocessing, and the necessity of knee point guidance. KPEA-N represents a variant of KPEA. During the environment selection of KPEA-N, solutions are selected based on the worse-elimination principle in VaEA, rather than on the penalty mechanism. This variant contains only preprocessing strategy and knee point guidance. KPEA-NR denotes a variant of KPEA that does not eliminate DRSs. KPEA-NK represents a variant of KPEA without a knee point-driven strategy. The IGD and HV results of KPEA, KPEA-N, KPEA-NR, and KPEA-NK on the MaF test suite are given in Tables 8 and 9, respectively.

Specifically, in terms of HV values, KPEA outperforms the other three variants on all test instances of the MaF4 and MaF11 problems. In terms of IGD values, KPEA outperforms KPEA-N, KPEA-NR, and KPEA-NK on 29 instances out of the 44 test instances. As shown in Table 8, for the MaF8 problem with distance minimization, there is no need to use preprocessing to eliminate DRSs. For problems with nonseparable Pareto fronts, the solutions obtained by KPEA-NR are very close to the final Pareto optimal front. In other words, the preprocessing strategy of KPEA may not be beneficial for optimization on such problems.

As shown in Table 9, for the MaF7 problem with concave-convex Pareto fronts, KPEA-NK performs significantly better than KPEA. This may be due to the fact that MaF7 is a multimodal problem that contains a large number of segmented Pareto-optimal fronts, which results in the KPEA easily falling into local Pareto optimality when driven by knee points.

It is important to note that on the MaF test suite, the performance of the KPEA-N is relatively the worst in both HV and IGD metrics. It is difficult for this variant to achieve relatively better performance on any problem. The adaptive penalty mechanism selects the individuals with the best convergence one by one, which achieves comprehensive coverage of the Pareto front. Compared to the worse-elimination principle in KPEA-N, the penalty mechanism in KPEA can better balance the convergence and diversity of the population.

However, both for IGD and HV values, KPEA has a significant advantage over the variant algorithms KPEA-N, KPEA-NR, and KPEA-NK in most test instances. Overall, the combination of the penalty mechanism, knee-point-driven strategy, and elimination of DRS strategy is more suitable for most problems.

5.5. Overall Performance Analysis

In this section, two statistical tests (i.e., the Wilcoxon rank-sum test and the Friedman test) are used to validate the comprehensive performance of KPEA in terms of convergence and diversity.

The statistical results based on the Wilcoxon rank-sum test are listed in Table 10. Specifically, in terms of the IGD metric, the proportion of the test instances where KPEA outperforms k-NSGAII, KnEA, Pi-MOEA, hpaEA, and VaEA with statistical significance is 50/80, 62/80, 69/80, 77/80, and 60/80, respectively. In terms of the HV metric, the proportion of the test instances where KPEA outperforms KPEA outperforms k-NSGAII, KnEA, Pi-MOEA, hpaEA, and VaEA with statistical significance is 74/80, 50/80, 46/80, 62/80, and 73/80, respectively. Moreover, the superiority of KPEA in terms of convergence and diversity is visually validated by Figure 14. These bar charts clearly indicate that KPEA achieves the best results of both HV and IGD indicators in most test instances.

Furthermore, in order to quantify the overall performance of each algorithm, a more effective statistical test, namely, the Friedman test, is also employed. The Friedman test ranks based on the IGD and HV results are presented in Table 11. KPEA ranks 1st in both HV and IGD metrics of the combined test suite. In addition, the bar chart in Figure 15 also intuitively shows that KPEA performs much better than the other five state-of-the-art algorithms in terms of the Friedman test ranking.

Both the results of the Wilcoxon rank-sum test and the average performance ranks indicate that KPEA has significant advantages in terms of convergence and diversity.

6. Discussions

In KPEA, a good balance between convergence and diversity is achieved through an adaptive switching mechanism. The knee point-driven selection mechanism increases selection pressure. The convergence and diversity of KPEA are maintained alternatively by the penalty mechanism. The experimental results in Section 5 demonstrate that KPEA can effectively solve optimization problems with different PFs.

In KPEA, when the number of nondominated solutions is less than the population size, knee points are first preserved by the selection mechanism, which increases selection pressure and improves convergence. Then, the remaining solutions are selected based on the principle of maximum-vector-angle-first, which maintains diversity. It is worth mentioning that both KnEA and k-NSGAII are knee point-driven algorithms that use adaptive neighborhood strategies to identify knee points. In addition, Pi-MOEA uses the average ranking method to identify pivot solutions within the neighborhood, which is similar to the strategy of identifying knee points. However, in KnEA and Pi-MOEA, the remaining nondominated solutions are selected based on the distance of each solution to the hyperplane. Therefore, this selection strategy emphasizes convergence too much but neglects the importance of diversity. Similarly, k-NSGAII uses parameter to control the width of the neighborhood, which leads to difficulty in maintaining satisfactory diversity.

Compared to the worse-elimination principle in VaEA, the penalty mechanism in KPEA can better balance the convergence and diversity of the population. However, the penalty mechanism uses a fixed neighborhood range to penalize solutions, which may be not suitable for addressing the problem with mixed PFs.

It is a challenging task to set an appropriate neighborhood range for the solution. Because too high or too low thresholds will affect the performance of the algorithm. Estimating the shape of the PFs or adaptively adjusting the neighborhood range may be a feasible approach.

Furthermore, recent research has pointed out that DRSs can have a significant impact on the performance of MaOEAs. Currently, scholars have proposed various strategies to eliminate DRSs, such as the direct identification of DRSs and the indirect elimination of DRSs by identifying boundary solutions. However, some boundary solutions of the extremely convex Pareto front (ECPF) are often mistakenly identified as DRSs. Indirect elimination of DRSs also cannot guarantee the identification of all boundary solutions. In real-world applications, such as unmanned aerial vehicle route planning [66] and software product line configuration [67], it is crucial to preserve ECPF boundary solutions. Therefore, it is necessary to accurately identify DRSs on the ECPF.

7. Conclusions

As the number of objectives increases, the performance of algorithms based on the Pareto dominance is severely affected. Recently, the angle-based selection strategies have been demonstrated to effectively address this problem. Motivated by this, this paper proposes a knee point-driven many-objective evolutionary algorithm with adaptive switching mechanism (KPEA). KPEA incorporates the knee points and the weighted distance into its mating strategy to increase the probability of generating excellent offspring. And an interquartile range method is introduced to detect and eliminate DRSs in the population. Moreover, in the environmental selection, an adaptive switching mechanism between angle-based selection and penalty is designed to balance convergence and diversity.

The performance of KPEA is compared with five state-of-the-art algorithms, namely, k-NSGAII, KnEA, Pi-MOEA, hpaEA, and VaEA, on WFG and MaF problems. The experimental results indicate that, in terms of IGD and HV, KPEA outperforms its competitors in most test instances. Additionally, this paper also discusses the sensitivity of parameter and provides recommendations for parameter settings.

KPEA still has some issues that require further research. First, when dealing with MaOPs with complex PFs, the neighborhood range of solutions will affect the coverage of the population on the Pareto front. Adaptively adjusting the neighborhood range during the evolutionary process is a feasible research direction. In addition, the DRS elimination strategy may face a dilemma between eliminating DRSs and preserving boundary solutions on some problems. Future research will focus on designing an effective strategy to differentiate between DRSs and boundary solutions.

Data Availability

The operational data used to support the findings of this study are included in the article.

Conflicts of Interest

The author states that this research paper does not have any known competitive economic interests or personal relationships.

Acknowledgments

This research was supported by the Innovation Project of the Engineering Research Center of Integration and Application of Digital Learning Technology (No. 1221003).