Abstract

Traditionally, in a multiobjective optimization problem, the aim is to find the set of optimal solutions, the Pareto front, which provides the decision-maker with a better understanding of the problem. This results in a more knowledgeable decision. However, multimodal solutions and nearly optimal solutions are ignored, although their consideration may be useful for the decision-maker. In particular, there are some of these solutions which we consider specially interesting, namely, the ones that have distinct characteristics from those which dominate them (i.e., the solutions that are not dominated in their neighborhood). We call these solutions potentially useful solutions. In this work, a new genetic algorithm called nevMOGA is presented, which provides not only the optimal solutions but also the multimodal and nearly optimal solutions nondominated in their neighborhood. This means that nevMOGA is able to supply additional and potentially useful solutions for the decision-making stage. This is its main advantage. In order to assess its performance, nevMOGA is tested on two benchmarks and compared with two other optimization algorithms (random and exhaustive searches). Finally, as an example of application, nevMOGA is used in an engineering problem to optimally adjust the parameters of two PI controllers that operate a plant.

1. Introduction

In industry, there are many situations where a design problem turns into an optimization problem. Moreover, it is usual that these problems have conflicting objectives, which generates a multiobjective optimization problem (MOP) ([13]). A MOP basically consists of three stages [4]: its definition, the optimization problem (search), and the multicriteria decision-making (MCDM). In an a priori multiobjective approach [5], the stages of optimization and decision-making are carried out at the same time, which results in a single solution, since the desired preferences are defined beforehand. On the contrary, in an a posteriori approach [6], the search does not supply a single solution but a set of optimal solutions (Pareto optimal solutions). This procedure is more time-consuming, but it gives the designer information about different solutions and provides them with a better understanding of the problem, which will allow them to make a well-informed decision in the decision-making stage. On the other hand, an optimization problem may have multimodal solutions ([79]) or nearly optimal solutions ([10, 11]) that are useful for the designer. These alternative solutions, which have a similar or even the same performance as the optimal solutions, are ignored in a traditional multiobjective approach, although they offer interesting and useful information to the designer. In order not to lose this valuable information, there must be taken into account not only the optimal solutions but also the nearly optimal ones ([12]).

Including the multimodal and nearly optimal solutions in the decision-making scenario increases the number of alternatives available to the designer. However, this poses two issues. On the one hand, the number of solutions will increase considerably and this may slow the algorithm to the point of becoming inoperative. On the other hand, having a huge amount of alternatives to choose makes the decision-making process more complicated. For this reason, it is important in most cases to reduce, in some way, the number of solutions. The way in which this reduction is performed will depend on the criterion or approach that the designer decides to adopt, which, in turn, depends on the application or the context. The existing algorithms usually perform this reduction by means of a discretization of the objective space. But this discretization is carried out without taking into account any consideration of the location of the solutions in the parameter space. The result of proceeding in this way is that nearly optimal solutions which have distinct characteristics could be neglected. Despite this drawback, this method is suitable for some applications.

One possible approach, however, and the one that we assume here, is the following: (1) finding a manageable number of solutions (so that the two issues mentioned above can be avoided) but, at the same time, (2) without neglecting the existing diversity in the characteristics of the nearly optimal solutions. We think that this approach is common in many applications and the existing algorithms do not serve in this case, precisely because they do not pay attention to the characteristics of the solutions. This is why, in this work, we have developed an algorithm to provide the designer with those solutions, among all the nearly optimal solutions, that fulfill that particular criterion, namely, the nondominated in their neighborhood. These solutions are the information that our designer finds more relevant. For example, a designer adopting this approach will not want to be given two nearly optimal solutions which have similar characteristics (neighboring solutions), if one of them is dominated by the other (i.e., worse for at least one of the objectives and not better for the rest), as he or she will logically choose the nondominated one. However, if these two solutions have significantly different characteristics (i.e., they are nonneighboring solutions), then both of them will be interesting for the designer. If the designer has these solutions available, they can analyze them a posteriori (for example, by including new indicators or by considering the physical sense of the solutions), in order to decide which one is the most suitable. From now on, we will call these solutions potentially useful solutions. This does not mean that the rest of nearly optimal solutions are useless, but simply that they are dispensable for a designer who has adopted the approach specified before. These potentially useful solutions are the ones that according to his or her criterion gives them the most relevant information, the information that they want to get.

An example where the specified approach could be the chosen one and, therefore, it is desirable to find the nearly optimal solutions nondominated in their neighborhood, is when one of the objectives cannot be included in the optimization process due to its high computational cost. In this case, it is possible to exclude that objective from the optimization process and to obtain the set of optimal solutions for the rest of objectives. When the optimization process has finished, the excluded objective can be evaluated (only at the optimal solutions previously obtained) and incorporated into the decision-making process. The problem here (if a classical MOP approach is followed) is that, since the excluded objective is ignored in the search, there will be interesting solutions (those with a good performance for the excluded objective and with similar performance for the rest of objectives) which will be missed. The new approach that we are proposing in this paper finds them. Another situation where it may also be interesting to examine nearly optimal solutions nondominated in their neighborhood happens when several objectives are aggregated, which is a way of defining a priori preferences. When a MOP has a lot of objectives, the decision-making stage becomes too complicated. That is why sometimes several objectives are aggregated in groups (for example, performance, cost, and robustness), in order to simplify the problem. However, this approach could miss valuable solutions for the designer. This loss of information can largely be avoided by finding the nearly optimal solutions nondominated in their neighborhood.

So, finding the nearly optimal solutions nondominated in their neighborhood solves the lack of information that typically appears in a classical multiobjective approach. Moreover, taking into account these solutions has two additional advantages: (1) it enables the designer to reassess the design objectives—if there are nearly optimal solutions which display certain characteristics that are absent in the optimal ones, this could mean a poor formulation of the objectives—and (2) it detects the existence of overparameterization (i.e., when the number of parameters is higher than needed, which could result in a man-made multimodality).

In this work, a new algorithm is presented, nevMOGA, which is aimed at finding the sets of optimal solutions and nearly optimal solutions nondominated in their neighborhood. There exist several algorithms in the literature that consider nearly optimal solutions as well ([13, 14]). However, they use these solutions simply as a means to compute an approximation of the Pareto front, whereas nevMOGA discriminates them and returns some of them (the nondominated in their neighborhood) as its outcome. nevMOGA is based on an algorithm called evMOGA, which is described in [15]. This new algorithm includes additional parameters to define when a solution is considered to be a nearly optimal solution and the size of a neighborhood.

2. Background

In this section, the concepts of multiobjective optimization problem and Pareto set are formally introduced, and the problem which we are dealing with is graphically explained (i.e., the problem of finding optimal and nearly optimal solutions nondominated in their neighborhood).

A multiobjective optimization problem (A maximization problem can be converted into a minimization one. For each of the objectives that have to be maximized, the transformation: can be applied.) can be stated as follows: where is defined as a decision vector in the domain and : is defined as the vector of objective functions . and are the lower and upper bounds of each component of .

Definition 1 (dominance [16]). A decision vector is dominated by any other decision vector if for all and for at least one , . This is denoted as

Definition 2 (Pareto set). The Pareto set (denoted by ) is the set of solutions in which are not dominated by another solution in :

Definition 3 (Pareto front). Given a set of Pareto optimal solutions , the Pareto front is defined as

Figure 1 shows an example to clarify what we mean by optimal solutions, nearly optimal solutions, and nearly optimal solutions nondominated in their neighborhood in a monoobjective (Figure 1(a)) and a multiobjective (Figure 1(b)) problem. Let us consider the monoobjective case. In theory, all the nearly optimal solutions (grey lines) could be interesting for the decision-maker. However, considering all these solutions makes the decision-making stage too difficult. Moreover, many of these solutions () will have very similar characteristics to the optimal one (), since they are close to it in the parameter space. This, in addition to the fact that they have a worse performance than the optimal one, leads us to consider them, under the criterion assumed, less relevant and in some cases dispensable. The rest of them () will have similar characteristics to , which is the best in and, therefore (always under the specified criterion), they could be dispensable. Consequently, the solutions which provide more relevant information (potentially useful solutions) are and , i.e., the optimal solution and the nearly optimal solution nondominated in its neighborhood. Likewise, in the multiobjective case, all the nearly optimal solutions (grey area) could be considered but, again, taking into account all of them would significantly complicate the decision-making stage while most of these solutions do not add much relevant information to the process. So, the sets (optimal solutions) and (nearly optimal solutions nondominated in their neighborhood) provide the most valuable information to the designer without unnecessarily complicating the decision-making process.

An example of the usefulness of finding the nearly optimal solutions nondominated in their neighborhood is when one of the objectives cannot be included in the optimization process due to its high computational cost. The usual way to cope with this difficulty consists of solving the MOP for the rest of objectives and then computing the excluded one only for the optimal solutions. The problem with that (if a classical MOP approach is followed) is that there will be nearly optimal solutions with an outstanding performance with respect to the excluded objective (and therefore worth considering) that will be ignored. On the other hand, if one is not willing to miss all those interesting solutions and decides to take into account all the nearly optimal solutions, then the time-consuming objective will be evaluated at a vast number of solutions, many of which are not even worth considering, as they will have similar characteristics to the optimal solutions (because they are neighboring solutions). In summary, approaching this kind of problems with a classical strategy leads to either neglecting potentially useful solutions or having to evaluate a time-expensive objective at many dispensable solutions, whose inclusion in the decision-making stage will, furthermore, unnecessarily complicate the decision. With our new approach, these problems can be overcome. Let us look at it with a concrete example.

Let us suppose a MOP with three objectives to minimize. Table 1 shows five possible solutions to the problem ( to ). In a classical MOP, all these solutions belong to the Pareto set (except either or , as they are multimodal solutions and, therefore, one of them would be discarded). Now, assume that the objective is computationally impracticable, i.e., its incorporation into the optimization process is impossible. For this reason, the search stage has to be solved for the first two objectives and then will be evaluated only at the solutions previously obtained. First, note that in a classical MOP, only the Pareto front is obtained, so either or (multimodal solutions) and would be discarded, although , , and seem to be good solutions, since they are nearly optimal solutions in different neighborhoods and thus worth considering.

With our new approach, , , and would also be found, because they are nondominated in their neighborhood; in other words, they have significantly different parameter values. would be discarded since it is dominated by and they are neighboring solutions, i.e., they are expected to have similar characteristic. has a better performance in than ; however, this improvement is insignificant (if the neighborhoods are well defined) as they are very similar solutions (neighboring solutions). Through this new procedure, from all the neighboring solutions, only those with the best performance over the objectives considered in the optimization process (in this example, and ) would be selected. The solution displays a better value for than and (which are in different neighborhoods) but worse values for and (this is why it is a nearly optimal solution) and, therefore, in a classical MOP, it would be discarded, although its analysis may be interesting for the decision-maker. Logically, would be ruled out since it is much worse for and , although it has a better value for . With the proposed approach, is only evaluated at , , and but not at ( is preferred instead) and (which is discarded, since it is not a nearly optimal solution) and, consequently, the computational burden is alleviated. Therefore, thanks to this new approach; it is possible to reduce the computational cost while obtaining the potentially useful solutions for the designer.

Another example that shows the usefulness of obtaining also the nearly optimal solutions nondominated in their neighborhood can be found when several objectives are aggregated, which is a common procedure when there are a great number of them. Let us suppose that a MOP originally has four objectives to minimize. Let be to (Table 2) the four solutions to this MOP. In a classical MOP, all these solutions would be selected to form the Pareto set.

If the designer wanted to reduce the number of objectives, a possibility would be to add and to obtain () and and to obtain (). But now, only or (multimodal solutions) would be the optimal solution (in a classical MOP). Let us suppose that the algorithm chooses and discards . In this case, and are discarded, although they are interesting solutions for the designer. However, if the algorithm keeps also the nearly optimal solutions nondominated in their neighborhood, and would not be excluded with the aggregated objectives. Finally, presents a poor performance for both and (clearly dominated) and, therefore, it does not seem to be a good solution.

In conclusion, when the objectives are aggregated, a classical MOP finds only one solution from the three that are interesting and the other two are missed. Therefore, finding the nearly optimal solutions nondominated in their neighborhood allows the designer to aggregate objectives (and, in this way, simplifying the decision-making stage) without losing potentially useful solution.

In summary, as seen in the examples, the multimodal solutions and the nearly optimal solutions which are significantly different from those which dominate them are solutions that can be potentially useful for the designer during the decision-making stage. This additional information is not provided to the decision-maker when these kinds of MOPs are approached in a classical manner.

3. Materials and Methods

In this section, we present in detail a novel algorithm, nevMOGA, which is the main contribution of this work, as well as the metric and benchmarks that have been used to demonstrate its correct functioning. Let us begin with the new set of interest and its discretization, which is performed by nevMOGA.

3.1. New Set of Interest

The nevMOGA algorithm is aimed at finding not only the optimal solutions but also the nearly optimal solutions nondominated in their neighborhood. In the following, both sets of solutions are defined.

Definition 4 (-dominance [17]). Define as the maximum acceptable performance degradation. A decision vector is -dominated by another decision vector if for all and for at least one , . This is denoted by .

Definition 5 (-Pareto set [18]). The -Pareto set (denoted by ) is the set of solutions in which are not -dominated by another solution in :

Definition 6 (neighborhood). Define as the maximum distance between neighboring solutions. Two decision vectors and are neighboring solutions () if for all .

Definition 7 (-dominance). A decision vector is -dominated by another decision vector if they are neighboring solutions (Definition 6) and . This is denoted by .

Definition 8 (-Pareto set). The -Pareto set (denoted by ) is the set of solutions of which are not -dominated by another solution in :

The set (optimal solutions and nearly optimal solutions nondominated in their neighborhood) is the new set of interest, i.e., what nevMOGA has to find. In order to clarify its nature, we will use a graphical example (see Figure 2). In this example, the set of interest is formed by the union of and . is a set of solutions nondominated in their neighborhood which belong to the set . is a set of solutions nondominated in their neighborhood which do not belong to the set . is not a nearly optimal solution, and, therefore, it will be discarded since it does not have the desirable performance (). is dominated (i.e., dominated by a neighboring solution) by and, as a result, this alternative provides less relevant information and, for this reason, is discarded. is dominated by nonneighboring solutions but not by neighboring solutions, so it belongs to , and the same can be said of any solution in . belongs to the optimal set . In summary, nevMOGA searches for the set , in our example, .

The set clearly depends on the values given to the new parameters (Definition 4) and (Definition 6). Given that in most problems, the objectives have a physical sense, it is possible to decide a priori how much performance we are willing to lose and to define consequently. Similarly, when the decision variables have a physical sense, the parameter can also be set intuitively. By setting , the designer establishes to what extent the two solutions are considered “similar”; in other words, setting defines the neighborhood.

If the decision variables do not have a physical sense and it is not evident how to choose , it is still possible to follow a simple procedure (see Figure 3) which will help us to do it if has been previously set. First, a reference solution is chosen. Second, in the objective space, a rectangle is defined with the center at and twice in width and in height. Then, starting from (in the search space), the value of each variable decision is increased and decreased independently, until the solution leaves the rectangle which was previously defined. Finally, each element of the vector neighborhood is set to the minimum distance between and the first solutions (one for each direction) which left the mentioned rectangle ( and in Figure 3). Through this procedure, we manage to quantify what excursion is needed in the search space in order to get significant changes in the objective space and, in this way, we are able to define when no other criterion is available.

3.2. Discretization of the New Set of Interest

Another issue is that may contain infinite solutions, so it is necessary to discretize it and to obtain a finite set of solutions () which are adequately distributed. In order to carry out this task, the concepts of box, _box, and their associated dominances have to be defined beforehand.

Definition 9 (box [19]). Let , with for all . Given a decision vector , is defined as the vector , where for , and where and are the maximum and minimum values of .

Definition 10 (box dominance [19]). Given two decision vectors and whose boxes are and , respectively, is said to box dominate , , if for all and for at least one , .

Definition 11 (-box dominance). Given the neighboring decision vectors and whose boxes are and , respectively, is said to -box dominate , , if .

The algorithm only keeps non--box-dominated solutions. A box cannot contain two neighboring solutions. When a neighboring solution is found in the same box, the nearest solution to the ideal corner (the lower left one) will be chosen (see Figure 4, and ) and the rest will be discarded. Figure 4 shows how the algorithm performs the discretization. is -box dominated (i.e., box dominated by a neighboring solution) by and, therefore, it is discarded. is box dominated but not -box dominated (since and are nonneighboring solutions), so it is an alternative, as it is a nearly optimal solution. , , and are in the same box. Two neighboring solutions cannot be in the same box, so either or has to be eliminated. In this case, is eliminated because it has a greater distance to the ideal corner than (). is a good alternative because it is a nearly optimal solution, it is not -box dominated and it does not belong to either the neighborhood of or .

The algorithm searches a discretization of the set of interest. This new discrete set is . The maximum number of solutions (in the worst case) of is (see 7). Similarly, and are the discrete sets of and , respectively.

3.3. Description of nevMOGA

The algorithm nevMOGA is based on the algorithm evMOGA [15]. The main difference between them is that nevMOGA computes an additional population, namely, the set of nearly optimal solutions nondominated in their neighborhood (subfront()). Its main characteristics are the following: (i)It is an elitist algorithm with two archives where the individuals of and are stored. It has also an additional, auxiliary population formed by the new individuals which are created during the process(ii)It is a real-coded algorithm and uses crossover (extended intermediate crossover) and mutation (random Gaussian distribution) operators

The algorithm manages four populations (see Figure 5): (1): by means of the population , the search space is explored, with the aims of obtaining the solutions of and having diversity in the set of solutions. The number of individuals in this population is constant and equal to (2)Front is the archive where is stored, i.e., a discrete approximation of the Pareto front. The size of this population varies but is always less than or equal to a given maximum size which depends on the number of boxes previously defined by the user(3)Subfront is the archive where is stored, i.e., the nearly optimal solutions nondominated in their neighborhood. Its size is variable but bounded, depending on the number of boxes(4) is an auxiliary population where the new individuals generated by the algorithm in each iteration are stored. The number of individuals of this population is , which must be multiple of 4

Now that the archives front and subfront have been presented, it is possible, with the help of the prior definitions, to establish the conditions that a solution must fulfill in order to enter them. This is the aim of the following two additional definitions:

Definition 12 (inclusion of in front). Given a solution and the archive front, will be included in front if and only if where is the distance from to the ideal corner (lower left) of the box (see Figure 3, and ). Additionally, if is included in front, then all the solutions that fulfill the following condition will be eliminated from front:

Then, it will be determined whether any of these solutions which have been eliminated from front is included in subfront or not, by applying Definition 13. Furthermore, all the solutions that fulfill the following condition will be eliminated from subfront:

Definition 13 (inclusion of in subfront). Given a solution such that , will be included in subfront if and only if

Additionally, if is included in subfront, then all the solutions that fulfill the following condition will be eliminated from subfront:

In order to maintain the diversity of solutions in , it is necessary that the population be diverse too. For this, is permanently ordered by using the niche count, which is an indicator of how densely populated a solution is in (Definition 14). In this way, during the processes of selection and substitution, the solutions in the less populated areas have a higher probability of being chosen to generate new individuals, whereas those in the more populated areas have a higher probability of being substituted.

Definition 14 (niche count [20]). where is the sharing function, which gives a measure of how similar two elements in a population are and is defined by

The algorithm pseudocode is the following:

1: t:=0;
2: Front (t):= ;
3: Sub-Front (t):= ;
4: Create initial population P (t) at random
5: Calculate () P (t)
6: Order population P (t) according to the niche count               Definition14
7: Inclusion of the individuals of P (t) in Front (t)               using Algorithm2
8: Inclusion of the individuals of P (t) Front (t) in Sub-Front (t)         using Algorithm3
9: for t 1:Number of iterations
10:  Create population G (t)                       using Algorithm4
11:  Calculate () G (t)
12:  Inclusion of the individuals of G (t) in Front (t)               using Algorithm2
13:  Inclusion of the individuals of G (t) Front (t) in Sub-Front (t)         using Algorithm3
14:  Update P (t) with the individuals of G (t)               using Algorithm5
15:  Order population P (t)
16: end for

Lines 3, 6, 8, 10, 13, 14, and 15 of the previous pseudocode constitute the changes which have been added to evMOGA, the algorithm on which ours is based. Next, the most important steps in nevMOGA are explained: (i)Lines 7 and 12 (Algorithm 2). Here, it is determined whether a new individual is included in or not. For this, the individual has to fulfill Definition 12. If the new individual is finally included in front, then it is analyzed whether its inclusion eliminates other individuals from front (in which case it will be established whether they must be included in subfront) and from subfront(ii)Lines 8 and 13 (Algorithm 3). Here, it is determined whether a new individual which was not included in front has to be included in subfront or not. For this, the individual has to fulfill Definition 14. If the new individual is included in subfront, then it is analyzed whether its inclusion eliminates other individuals from subfront(iii)Line 10 (Algorithm 4). The population is created by using the following procedure: (1)Two individuals are chosen at random, one from front () (where any individual has the same probability of being chosen) and another from () (where each one has a different probability of being chosen). In effect, the solutions of are ordered according to their niche count (from least to greatest) and its individual is chosen according to an exponential distribution, which makes it more likely for a solution with a lower niche count to be chosen. This procedure favors the uniform distribution of the solutions(2)A random number is generated. If (probability of crossover/mutation), then a crossover will be performed (step 3); otherwise, a mutation will be performed (step 4)(3) and are crossed by means of an extended intermediate crossover operator, which produces two new individuals and : The parameter is a random value uniformly distributed which belongs to interval , and is a parameter which is adjusted by using an exponential decreasing function, as in simulated annealing [19]: Figure 6 shows from to , where and .(4) and are mutated by using a random mutation with Gaussian distribution: where the variances and are expressed as a percentage of . These variances are calculated by a function which is similar to the one previously used for the parameter : (5) and are included in These five steps are repeated again from step one but now with a solution from subfront instead of one from front. The whole process is executed again and again until is full. In order to set the parameters which appear in the equations of the steps 3 and 4, the default values suggested by [19] for the original algorithm (evMOGA) are taken.(iv)Line 14 (Algorithm 5). In the updating of , it is determined whether a new individual from () substitutes any other existing individual in or not. The search for the individual to be substituted () starts from an individual chosen by an exponential distribution which is the inverse of the one used for the choice of in the creation of (line 10, Algorithm 4) and, therefore, now the more populated solutions will be more likely to be chosen. This is aimed at achieving a more uniform distribution of the solutions. During this updating, there may occur three distinct cases: (1)The new individual is dominated by some member of front. In this case, an individual dominated by will be searched to be substituted(2)The new individual is not dominated by any member of front, and there exists an individual dominated by which is dominated by some member of front. In this case, it will be substituted(3)The new individual is not dominated by any member of front, and there does not exist an individual dominated by which is not dominated by any member of front. In this case, the initial solution is chosen. If and are neighboring solutions, a solution dominated by will be searched to be substituted. If they are not, a solution dominated by will be searched to be substituted. In this last case, if that solution is not found, will be randomly substituted by a solution from the neighborhood of

1: if then
2:  return
3: else if then
4:  return
5: end if
6: Include in Front (t)
7: if then           Remove non-optimal solutions
8:  Remove from Front (t)
9: else if then
10:  Remove from Front (t)
11: end if
12: if then          Remove non-nearly-optimal solutions
13  Remove from Sub-Front (t)
14: end if
1: if then
2:  return
3: else if then
4:  return
5: else if then
6:  return
7: end if
8: Include in Sub-Front (t)
9: if then      Remove non-nearly-optimal solutions
10:  Remove from Sub-Front (t)
11: else if then
12:  Remove from Sub-Front (t) then
13: end if
1: Choose Front (t) Random selection
2: Choose P (t) According to an exponential distribution ()
3: u random(1)
4: if u> then is the probability of crossover and mutation
5:   and are obtained by crossing over and
6: else
7:   and are obtained by mutating and
8: end if
9: Choose Sub-Front (t) Random selection
10: Choose P (t) According to an exponential distribution ()
11: u random(1)
12: if u> then
13:   and are obtained by crossing over and
14: else
15:   and are obtained by mutating and
16: end if
1: Choose P (t)        According to an exponential distribution ()
2: if Front (t) then
3:  if P (t): Front (t) then The search for starts from
4:   
5:   return
6:  else
7:   
8:   if then
9:    if then            The search for starts from
10:     
11:     return
12:    end if
13:   else
14:    if then
15:     
16:     return
17:   end if
18:  end if
19: end if
20:   =Random()              Random selection of an individual from
21:  
22: else
23:  if P (t): then             The search for starts from
24:   
25:  end if
26: end if
3.4. Performance Metrics

Next, we present one metric which will serve to assess the performance of the different algorithms to be compared. This metric measures both the convergence and the diversity between two sets (the outcome set and the target set). The outcome set of each algorithm is the set of solutions that it returns, whereas the target set results from discretizing the search space with a fine grain. We use this metric to measure the performance of the algorithms in the objective space (convergence toward the front) and in the parameter space (diversity of solutions).

3.4.1. Averaged Distance Hausdorff

A single indicator is used to measure the convergence and diversity between two sets (averaged distance Hausdorff , [21]). The sets and have and solutions, respectively. The is calculated by where

This indicator is the average of the Hausdorff distance, and we use it here in order to measure the performance between an outcome set and a target set . This performance is measured both in the objective space and in the parameter space. In this way, it is possible to measure convergence toward the front and diversity of solutions. We use . It is desirable that this indicator has the smallest possible value. Ideally, when , then .

3.5. Benchmarks

Our algorithm nevMOGA has been tested on two distinct benchmarks. In this section, we describe them.

3.5.1. Benchmark 1

The first benchmark considered in this work corresponds to an academic example [18] and is stated as follows: where subject to

Setting , , and , this MOP has one global Pareto set: and eight local Pareto sets:

3.5.2. Benchmark 2

The second benchmark is an adaptation of the modified Rastrigin [22]. This benchmark has been modified in order to turn it into a MOP with one optimal set and nearly optimal solutions lying in different neighborhoods. This is its formulation: where and and subject to

4. Results and Discussion

In this section, we present the results of nevMOGA on the two benchmarks introduced and on a control design problem. These results will be analyzed on the basis of the metric which was introduced in Section 3.4. In addition, we compare nevMOGA with two other algorithms, namely, random and exhaustive searches. In order to make this comparison possible, all three algorithms evaluate the objective functions the same number of times. Although there do exist algorithms which search the set of nearly optimal solutions ([17, 18, 23]), none of them has the same set of interest as nevMOGA (). For this reason, nevMOGA cannot directly be compared with any of them. In order to obtain statistical results, each algorithm is tested 50 times on each benchmark. For the exhaustive search case, the evaluated points are uniformly distributed over the search space, forming a hypergrid. In order for it to obtain different results, the hypergrid is slightly lifted each time.

4.1. Benchmark 1

In this benchmark, since the global Pareto set and the local Pareto sets are known (Section 3.5.1), we can set the parameters and more easily. Here, we set so that the eight sets described in (25) are regarded as nearly optimal solutions. We get by applying the procedure described in Section 3.1 with a reference solution and the parameter previously set, resulting in . Once these parameters have been determined, the set to be found (target set , see Figure 7) is

Figure 8 and Table 3 show the statistical results of the three algorithms (each one is tested 50 times on this benchmark) that have been compared (nevMOGA, random search, and exhaustive search) using the metric . Each algorithm evaluates the objective function 5000 times in each test run. The outcome sets of the three algorithms are compared with the target set . In Figure 8(a), there are the results against , whereas in Figure 8(b) against , using boxplots and density estimation. The values yielded by nevMOGA for the metric in the objective space and in the parameter space are good (near 0) and better than those provided by random and exhaustive searches. For the random search, the approximation quality to the front varies enormously. For the exhaustive search, the results display less variability, although it performs worse than nevMOGA in both spaces. Moreover, a multiple comparison test ([24, 25]) leads to the conclusion that the results of the three algorithms are significantly different with respect to the metric in both spaces. For these reasons, it can be concluded that nevMOGA significantly outperforms random and exhaustive searches on benchmark 1.

In order to graphically compare the three algorithms, a representative solution (from the 50 test runs) of each one has been plotted in Figure 9. This representative solution has been chosen as the one whose values of and are closest to the average values of that metric in both spaces. The outcome set obtained by nevMOGA has 70 solutions, whereas those obtained by random and exhaustive searches have 78 and 48 solutions, respectively. The great number of solutions found by nevMOGA in addition to the results shown in Figure 9 proofs that nevMOGA is capable of characterizing the target set better than its competitors (random and exhaustive searches) on benchmark 1. This result, plus the fact that nevMOGA achieves lower values of the metric , leads us to conclude that nevMOGA accomplishes a better general performance on this benchmark.

This benchmark has also been solved by [18], where the set is characterized. The algorithm tested in that work is called -MOEA. A comparison of the results of nevMOGA with those of -MOEA shows that nevMOGA obtains a significantly smaller number of solutions (approximately four times smaller) without missing any neighborhood containing nearly optimal solutions. This was our goal, namely, finding only those solutions that provide the most relevant information, and this with two aims: (1) not to slow down the algorithm and (2) to simplify the decision-making stage.

4.2. Benchmark 2

The statement of this second optimization problem was presented in Section 3.5.2. First, and with the aim of setting the two parameters of nevMOGA ( and ), a tentative search for solutions is carried out within the defined range. This leads us to choose . Once has been set, we take a reference solution and follow the procedure detailed in Section 3.1 to obtain . This results in . The selection of and leads to a particular target set , which is shown in Figure 10. As can be seen in the figure, is dominated by the set but they are in distinct neighborhoods. The set does not have any solution dominated in its neighborhood and, therefore, it is part of the target set .

Figure 11 is similar to Figure 8 but for benchmark 2. Each algorithm is run 50 times and can evaluate the objective function 5000 times in each test run. Again, nevMOGA achieves better results than random and exhaustive searches with regard to the metric , since it reaches values which are closer to zero. As we did in benchmark 1, we have also performed a multiple comparison test on these results. This comparison shows that the results obtained by the three search strategies are significantly different in both spaces. For these reasons, it can be concluded that nevMOGA significantly outperforms random and exhaustive searches on benchmark 2.

Figure 12 shows a representative solution (chosen in the same way as we did in benchmark 1) of each algorithm (nevMOGA, random, and exhaustive searches). A glance at the figure is sufficient to show that nevMOGA accomplishes a better approximation to the target set . Moreover, the outcome set obtained by nevMOGA has 73 solutions, whereas those obtained by random and exhaustive searches have 62 and 48 solutions, respectively. These two last facts (a better approximation to the target set and a greater number of solutions) confirm the conclusion drawn before, when the results were statistically examined, namely, that nevMOGA outperforms both random and exhaustive searches.

4.3. Multiobjective Control Design Problem

In this section, we present the application of nevMOGA to a real design problem in the field of control engineering. The problem consists of optimizing the performance of two PI controllers, which operate a multivariable plant with two inputs and two outputs.

The model of the plant is [26] where and are the outputs of the plant, and and are the inputs (control actions). The time constants are all in seconds.

The control structure (also taken from [26]) is defined as a multiloop PI control which uses an off-diagonal pairing scheme, i.e., output is controlled by and by . That is to say, where and are the proportional gains, and are the integral time constants, and and are the errors, where and are the setpoints for and , respectively (see Figure 13).

The optimization problem is stated as follows: where subject to and where

We launch two different simulations, sequentially. In the first one, a unitary step is applied to , whereas is kept to zero. In the second one, the unitary step is now applied to and is kept to zero. In each simulation, the metric ISE (Integral Squared Error) is calculated for both outputs and . Our first objective is then defined as the sum of the values of the ISE metrics of each simulation. So, the objective measures the setpoint tracking performance of the control system. The second objective is calculated in a similar manner but using the control action signals instead of the errors. Therefore, measures the control effort. Note that each objective ( and ) can be seen as a sum of 4 subobjectives (see 32).

Once the optimization problem has been defined, the two parameters of nevMOGA ( and ) have to be set. It is important to choose these parameters correctly, especially . First of all, we have to decide what performance loss (with respect to the Pareto set) we are willing to accept (i.e., the parameter ), in other words, what we are still going to regard as a nearly optimal solution. Choosing a high value would lead to a great number of possible solutions, whereas a low value would provide very few. In this application example, although the design objectives do have a physical sense, it is not self-evident how to choose . In order to solve this difficulty, we have proceeded in the following way. First, we have taken the values of and corresponding to the control designed in [26] as a reference solution. Then, we have pinpointed that solution on the Pareto front (which we have taken from [27]). Finally, on the basis of that information, we have decided to choose . Note that there is an unavoidable subjective element concerning the way in which is chosen. Now, we have to assign a value to , which is the most critical parameter of the algorithm. On the one hand, choosing an excessively low value would make most solutions be nonneighboring solutions, which in turn would cause the algorithm to tend to obtain the whole set . On the other hand, choosing an excessively high value would result in many solutions being neighboring solutions, which would cause the algorithm to tend to obtain only the set . This is why it is crucial that be chosen carefully, supporting the choice with good reasons. Next, we describe how we have set . First, we choose a reference controller (taken from [26]). Second, starting from the solution corresponding to this reference controller, the value of each of its parameters is increased and decreased independently, until reaching significant changes in the system response (see Figure 3). In this way, we finally get ; these are the values from which the system response is significantly different from the one provided by the reference controller.

As mentioned before, the objective functions include several terms which measure the performance and the control effort of each control loop (as if there were 4 subobjectives added for each objective). For this reason, there may be solutions with the same objective value which are, however, completely different. This way of defining the objective functions is useful when the number of control loops is high, since it reduces the number of objectives and simplifies the decision-making stage. Yet, this approach has a shortcoming, namely, many interesting solutions may be neglected, because they may become nearly optimal solutions or multimodal solutions. Figure 14 shows the set of solutions found by nevMOGA (for a given discretization of the objective space), where , , and are the optimal solutions, and , , and are the nearly optimal solutions. In order to show the decision variables, we employ a visualization tool called level diagrams ([28, 29]). For the ordinate axis, we have used the 2-norm. For the graphical representation of the objectives, we have not used this tool though, since there are only two and, in this case, plotting them on a single graph, where they can directly be compared, is preferable. Let us draw attention to the decision variables and (Figures 14(b)). Note how nevMOGA has been able to identify new areas in the search space containing solutions which may be interesting for the decision-maker and that would have been neglected in a traditional optimization approach, where only the set of Pareto optimal solutions is characterized.

Figure 15 shows the system responses for the solutions (optimal solution) and (nearly optimal solution), which have very similar values of and (see Figure 14(a)). These two solutions lie in distinct neighborhoods, and their corresponding responses are different too. This should not cause surprise since their corresponding controllers are significantly dissimilar from each other.

A more careful comparison of these two transient responses shows that presents a much better response than with respect to when the step is applied to , but presents a better response than with respect to and (better in the sense that it displays less overshoot) when the step is applied to . Therefore, it is not clear at all which solution should be chosen. In these cases, the decision-maker will have to include new criteria or preferences, in order to make the decision. Anyway, note that the nearly optimal solution could ultimately be preferred over the optimal one . This evidences that it would not have been wise to discard in the search stage and proofs the convenience of finding the nearly optimal solutions besides the optimal ones. This situation is very common when the objective functions are defined as an aggregation of several subobjectives, as is the case here. Additionally, it is worth noting that if this MOP had been addressed with a traditional approach and eight objectives (one for each term in 32, without adding objectives)—which is the ideal approach if one wants to find all the optimal solutions, although impractical in the decision-making stage—then would not be dominated by and both and could belong to the Pareto front. This is an additional evidence of the relevance of , which nevMOGA has been able to identify.

Next, we assess the robustness of every solution found by the algorithm, that is to say, how well each controller maintains its performance when identification errors are contemplated. This analysis is typical in control design. In order to do that, first, we get 50 sets of plant parameter (gains, time constants, and time delays) by generating random values for each parameter within a range of 20%. Each of this sets represents a deviation from the nominal plant. Then, we evaluate the performance ( and ) of each controller for each one of those deviations. Including an additional objective which measures robustness and computing it in the optimization stage, along with the two other objectives, would have been very expensive computationally. This is why this robustness analysis is shifted to the decision-making stage. As an example of the results of this robustness analysis, Figure 16 shows the boundaries of the performance degradation (determined by the values of and when they are evaluated for each one of the 50 sets of plant parameters) experienced by two controllers ( and ). Note that (a nearly optimal solution) experiences less degradation than . This is a very valuable information and could make the decision-maker opt for instead of .

Let us look at one last proof of the convenience of considering the nearly optimal solutions. If we compare (optimal solution) and (nearly optimal solution), we see that they have a similar performance with respect to both and . However, has lower values (in absolute value) of the gains and . From a design perspective, this fact may be an advantage, since this solution will present less sensibility to any measurement noise in the outputs of the plant and, therefore, it could be preferred over the optimal solution.

5. Conclusions

In this paper, we have presented a new multiobjective genetic algorithm, nevMOGA, which is capable of characterizing not only the Pareto set but also the set of nearly optimal solutions nondominated in their neighborhood. We think that this algorithm will be useful for designers who want to get a manageable number of solutions without neglecting the existing diversity in the characteristics of the solutions. In comparison with other algorithms that only return the Pareto front, our algorithm provides, in addition, new alternatives to be analyzed in the decision-making stage. Given that these solutions have a very similar performance to the optimal ones and significantly different values in the parameter space, they will be regarded by a designer adopting the specified approach as solutions worth studying. This is why obtaining these solutions, which we have called potentially useful solutions, can be very valuable in many multiobjective optimization problems.

nevMOGA has been compared with two other algorithms (random search and exhaustive search) on two benchmark problems, in order to assess its performance. It has not been compared to any other because no genetic algorithm of similar characteristics (one which is capable of finding nearly optimal solutions nondominated in their neighborhood) is available in the literature. In order to perform this comparison, the metric has been employed, which measures the approximation quality of the outcome set of an algorithm to the target set (the set of solutions to be found) in the objective space and in the parameter space. In addition, a statistical analysis has been conducted. The results reveal the outperformance of nevMOGA in characterizing the set of optimal and nearly optimal solutions in both spaces.

nevMOGA has also been applied to a real design problem in the field of process control engineering. This design problem consists of optimally adjusting the parameters of two PI controllers ( and ) which operate a plant with two inputs and two outputs. The control performance is evaluated by two widely used, conflicting metrics: , which measures the setpoint tracking performance of the control system, and , which measures the control effort. Each of these objective functions has been formulated as a sum of four subobjectives (see 32), which means that eight different objectives have been incorporated in the problem definition. Dealing with eight independent objective functions would have complicated enormously the analysis of the Pareto front and, consequently, the decision-making stage. This is why the eight objectives have been aggregated to form only two objective functions, whose Pareto front can be plotted on a two-dimensional graph and, thus, easily analyzed.

Aggregating objectives has, however, a major drawback if the optimization algorithm searches only for the optimal solutions: the resulting set of Pareto optimal solutions will lack many solutions that may be interesting and useful for the designer and, in some cases, even preferable to the optimal ones. nevMOGA, as it searches for nearly optimal solutions nondominated in their neighborhood too, is capable of finding those valuable solutions which are neglegted by a classical optimization algorithm. This means that the decision-maker will have a greater number of alternatives, all of them worth considering. For instance, if we compare (a nearly optimal solution) and (an optimal solution), we see that they have similar performances () but their corresponding system responses are different, since they lie in distinct neighborhoods. The designer could choose the nearly optimal solution instead of the optimal solution , since the former achieves a better system response when the step is applied to the setpoint . A comparison between (optimal solution) and (nearly optimal solution) leads to the same conclusion: they have similar performances () but has lower values of and (the gains of the PI controllers) than and, therefore, will be less sensitive to the measurement noises in the outputs, which is a considerable advantage in some applications. This could make the designer prefer to .

For this same design problem, a robustness analysis has been performed on the optimal and nearly optimal solutions obtained in the optimization stage. The results of this analysis show how a nearly optimal solution () can be more robust than an optimal one () while providing a similar performance with respect to the objective function (). This is due to the fact that they are nonneighboring solutions and, therefore, likely to present distinct characteristics. If we had included robustness as a metric in the definition stage in the same way as we did with and , the computational cost would have increased drastically (around 50 times higher).

These examples demonstrate the pertinence of the criterion assumed and, consequently, the great convenience of taking into account the potentially useful solutions, that is to say, the nearly optimal solutions that are not dominated in their neighborhood. This approach is specially advisable when the MOP presents multimodal solutions, either because they are intrinsic to the problem or because they have arisen artificially as a result of aggregating several objectives into one single-objective function. Likewise, having the nearly optimal solutions available represents an advantage when there exist objectives which cannot be incorporated within the optimization process due to its high computational cost. When this happens, these objectives are excluded from the optimization stage and evaluated afterwards only for the optimal and nearly optimal solutions resulting from the search.

Data Availability

The data used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest

The authors declare that there is no conflict of interest regarding the publication of this paper.

Acknowledgments

This work was partially supported by the Ministerio de Economía y Competitividad (Spain) Grant numbers DPI2015-71443-R and FPU15/01652, by the local administration Generalitat Valenciana through the project GV/2017/029, and by the National Council of Scientific and Technological Development of Brazil (CNPq) through the grant PQ-2/304066/2016-8.