Loading [MathJax]/extensions/TeX/boldsymbol.js
Performance of Decomposition-Based Many-Objective Algorithms Strongly Depends on Pareto Front Shapes | IEEE Journals & Magazine | IEEE Xplore

Performance of Decomposition-Based Many-Objective Algorithms Strongly Depends on Pareto Front Shapes


Abstract:

Recently, a number of high performance many-objective evolutionary algorithms with systematically generated weight vectors have been proposed in the literature. Those alg...Show More

Abstract:

Recently, a number of high performance many-objective evolutionary algorithms with systematically generated weight vectors have been proposed in the literature. Those algorithms often show surprisingly good performance on widely used DTLZ and WFG test problems. The performance of those algorithms has continued to be improved. The aim of this paper is to show our concern that such a performance improvement race may lead to the overspecialization of developed algorithms for the frequently used many-objective test problems. In this paper, we first explain the DTLZ and WFG test problems. Next, we explain many-objective evolutionary algorithms characterized by the use of systematically generated weight vectors. Then we discuss the relation between the features of the test problems and the search mechanisms of weight vector-based algorithms such as multiobjective evolutionary algorithm based on decomposition (MOEA/D), nondominated sorting genetic algorithm III (NSGA-III), MOEA/dominance and decomposition (MOEA/DD), and θ-dominance based evolutionary algorithm (θ-DEA). Through computational experiments, we demonstrate that a slight change in the problem formulations of DTLZ and WFG deteriorates the performance of those algorithms. After explaining the reason for the performance deterioration, we discuss the necessity of more general test problems and more flexible algorithms.
Published in: IEEE Transactions on Evolutionary Computation ( Volume: 21, Issue: 2, April 2017)
Page(s): 169 - 190
Date of Publication: 12 July 2016

ISSN Information:

Funding Agency:


SECTION I.

Introduction

Recently the development of evolutionary multiobjective optimization (EMO) algorithms was discussed from the point of view of co-evolution with test problems in [1] where the Pareto front was compared with randomly generated initial solutions to explain the characteristic features of each test problem. For example, Fig. 1 shows a test problem in [2] used for performance evaluations of EMO algorithms in the mid-1990s. As we can see from Fig. 1, some initial solutions are very close to the Pareto front. This observation may explain why nonelitist EMO algorithms such as nondominated sorting genetic algorithm (NSGA) [3] and Niched Pareto genetic algorithm (NPGA) [4] with no strong convergence property were proposed in the mid-1990s. However, initial solutions of a test problem in Fig. 2 [5] are not close to the Pareto front. Thus we need strong convergence whereas no strong diversification is needed in Fig. 2 (since randomly generated initial solutions have a large diversity). These observations explain why elitist EMO algorithms such as strength Pareto evolutionary algorithm (SPEA) [6], NSGA-II [7], and SPEA2 [8] were proposed around 2000. In these algorithms, Pareto dominance was used as the main fitness evaluation criterion together with a secondary criterion for diversity maintenance.

Fig. 1. - Randomly generated 200 solutions and the Pareto front of a test problem in [2].
Fig. 1. - Randomly generated 200 solutions and the Pareto front of a test problem in [2].
Fig. 1.

Randomly generated 200 solutions and the Pareto front of a test problem in [2].

Fig. 2. - Randomly generated 200 solutions and the Pareto front of a two-objective ZDT1 problem [5].
Fig. 2. - Randomly generated 200 solutions and the Pareto front of a two-objective ZDT1 problem [5].
Fig. 2.

Randomly generated 200 solutions and the Pareto front of a two-objective ZDT1 problem [5].

In parallel with the increase in the popularity of the Pareto dominance-based EMO algorithms [6]–​[8], many-objective test problems called DTLZ [9] and WFG [10] were proposed as scalable test problems where the number of objectives can be arbitrarily specified. Multiobjective knapsack problems [6] were also generalized to many-objective test problems with up to 25 objectives [11], [12]. The DTLZ, WFG, and knapsack problems were repeatedly used for demonstrating difficulties of many-objective optimization for the Pareto dominance-based EMO algorithms [13]–​[16]. When an EMO algorithm is applied to a many-objective problem, almost all solutions in a population become nondominated with each other in very early generations (e.g., within ten generations) before they converge to the Pareto front. This means that the Pareto dominance-based selection pressure toward the Pareto front becomes very weak. As a result, convergence ability of Pareto dominance-based EMO algorithms is severely degraded by the increase in the number of objectives.

For improving the convergence ability for many-objective problems, various approaches have been proposed such as the modification of the Pareto dominance relation [17] and the introduction of an additional ranking mechanism [18]–​[21]. The use of a different fitness evaluation mechanism was also actively studied. Approaches in this direction can be classified into two categories. One is an indicator-based approach such as S metric selection evolutionary multiobjective optimization algorithm (SMS-EMOA) [22] and hypervolume estimation algorithm for multiobjective optimization (HypE) [23]. The advantage of EMO algorithms in this category is a sound theoretical support, which is based on a direct relation between many-objective optimization and indicator optimization. Their main difficulty is heavy computation load for fitness evaluation. The other category is characterized by the use of a scalarizing function for fitness evaluation such as multiple single objective Pareto sampling (MSOPS) [24] and multiobjective evolutionary algorithm based on decomposition (MOEA/D) [25]. These algorithms are referred to as decomposition-based or scalarizing function-based algorithms.

Whereas MOEA/D [25] was not originally proposed for many-objective problems, its high performance as a many-objective optimizer was observed [16], [26]. Recently, a number of new many-objective algorithms have been proposed using the framework of MOEA/D (e.g., improved decomposition-based evolutionary algorithm (I-DBEA) [27], MOEA/D-distance-based updating strategy (MOEA/D-DU) [28], Ensemble Fitness Ranking with a Ranking Restriction scheme (EFR-RR) [28], and \theta -DEA [29]). Some other algorithms (e.g., NSGA-III [30], MOEA/DD [31], and U-NSGA-III [32]) can be viewed as using hybrid mechanisms of the Pareto dominance-based fitness evaluation and the MOEA/D framework.

MOEA/D [25] searches for well-distributed solutions using systematically generated weight vectors. As an example, we show a set of 91 weight vectors for a three-objective problem in Fig. 3. In Fig. 4, we show an example of obtained solutions by MOEA/D-PBI with \theta ={\mathrm{ 5}} [25] for a three-objective DTLZ2 problem [9]. Well-distributed solutions were obtained in Fig. 4 using the uniformly distributed weight vectors in Fig. 3.

Fig. 3. - Example of weight vectors used in MOEA/D (91 weight vectors).
Fig. 3. - Example of weight vectors used in MOEA/D (91 weight vectors).
Fig. 3.

Example of weight vectors used in MOEA/D (91 weight vectors).

Fig. 4. - Example of obtained nondominated solutions by MOEA/D-PBI for a three-objective DTLZ2 problem using the 91 weight vectors in Fig. 3.
Fig. 4. - Example of obtained nondominated solutions by MOEA/D-PBI for a three-objective DTLZ2 problem using the 91 weight vectors in Fig. 3.
Fig. 4.

Example of obtained nondominated solutions by MOEA/D-PBI for a three-objective DTLZ2 problem using the 91 weight vectors in Fig. 3.

In Figs. 3 and 4, a one-to-one mapping was realized between the weight vectors in Fig. 3 and the obtained solutions in Fig. 4. However, this is not always the case. As examined in [33], the number of obtained nondominated solutions by MOEA/D is often much smaller than the number of weight vectors. This is because: 1) a single good solution can be shared by multiple weight vectors and 2) all solutions are not always nondominated. Recently proposed many-objective algorithms [27]–​[32] with the MOEA/D framework have mechanisms for improving both the convergence of solutions toward the Pareto front and their uniformity over the entire Pareto front. Their common feature is the use of systematically generated weight vectors. In those algorithms, reference points and/or reference lines are constructed using the weight vectors.

Surprisingly good results were reported by those weight vector-based algorithms on the DTLZ and WFG problems in the literature. For example, the average inverted generational distance (IGD) over 20 runs on a 15-objective DTLZ2 problem was reported in [29] as {\mathrm{ 1}}.{\mathrm{ 726}}\times {\mathrm{ 1}}0^{-{\mathrm{ 2}}} by NSGA-III, 1.133\times 10^{-2} by \theta -DEA, and 6.005\times 10^{-3} by MOEA/D-PBI. In Fig. 5, we show the reference points used for the IGD calculation in [29] and a set of obtained solutions by a single run of MOEA/D-PBI with \theta ={\mathrm{ 5}} (IGD is 5.590\times 10^{-3} in Fig. 5). The obtained solution set in Fig. 5(b) is almost the same as the reference point set in Fig. 5(a).

Fig. 5. - Reference points for the IGD calculation and a set of obtained solutions by MOEA/D-PBI with 
$\theta = 5$
 for a 15-objective DTLZ2 problem. (a) Reference points. (b) Obtained solutions by a single run of MOEA/D-PBI.
Fig. 5. - Reference points for the IGD calculation and a set of obtained solutions by MOEA/D-PBI with 
$\theta = 5$
 for a 15-objective DTLZ2 problem. (a) Reference points. (b) Obtained solutions by a single run of MOEA/D-PBI.
Fig. 5.

Reference points for the IGD calculation and a set of obtained solutions by MOEA/D-PBI with \theta = 5 for a 15-objective DTLZ2 problem. (a) Reference points. (b) Obtained solutions by a single run of MOEA/D-PBI.

Whereas difficulties of many-objective problems were repeatedly pointed out (e.g., see survey papers [12], [34], and [35], Fig. 5 and the reported results in [27]–​[32] may suggest that some DTLZ and WFG problems are not difficult. In this paper, we first examine why surprisingly good results were obtained for some DTLZ and WFG test problems. Then we show that a slight change in the DTLZ and WFG formulations degrades the performance of weight vector-based many-objective algorithms. Our experimental results suggest the overspecialization of those algorithms for the test problems.

This paper is organized as follows. In Section II, we briefly explain the DTLZ [9] and WFG [10] problems. We focus on the shape of the Pareto front of each test problem. In Section III, we explain a common search mechanism of weight vector-based algorithms such as MOEA/D-PBI [25], \theta -DEA [29], and NSGA-III [30]. We also discuss the relation between the shape of the Pareto front of each test problem and the common search mechanism of those algorithms. In Section IV, we demonstrate their high performance on the DTLZ and WFG problems. We also demonstrate severe performance deterioration by a slight change in the problem formulations of DTLZ and WFG: all objectives are multiplied by (−1). In Section V, we explain why the performance of the weight vector-based algorithms is sensitive to such a slight change. We also discuss the necessity of more general test problems and more flexible algorithms. Finally, we conclude this paper in Section VI.

SECTION II.

Many-Objective Test Problems

In general, an {M} -objective minimization problem with a decision vector \boldsymbol {x} and its feasible region \boldsymbol {X} is written as \begin{equation} {\mathrm{ Minimize}}~f_{1} ({ \boldsymbol {x}}), f_{2} ({ \boldsymbol {x}}), \ldots , f_{M} ({ \boldsymbol {x}})~{\mathrm{ subject~to}}~{ \boldsymbol {x}}\in { \boldsymbol {X}} \end{equation} View SourceRight-click on figure for MathML and additional features. where f_{i} ({ \boldsymbol {x}}) is the {i} th objective to be minimized (i = 1, 2, \ldots , M ). In the EMO community, multiobjective problems with four or more objectives (i.e., M\ge 4 ) are often referred to as many-objective problems [12], [34], [35].

Scalable many-objective test problems called DTLZ [9] and WFG [10] have been frequently used to evaluate many-objective algorithms in the literature. Table I summarizes many-objective test problems used for performance evaluation of recently proposed weight vector-based algorithms [27]–​[32]. For comparison, we also show multiobjective test problems used for evaluating MOEA/D in [25]. We can see from this table that DTLZ1-4 and WFG1-9 have often been used in the literature. In this section, we briefly explain those test problems.

TABLE I Many-Objective Test Problems Used in Each Study on Weight Vector-Based Many-Objective Evolutionary Algorithms
Table I- 
Many-Objective Test Problems Used in Each Study on Weight Vector-Based Many-Objective Evolutionary Algorithms
Table I- 
Many-Objective Test Problems Used in Each Study on Weight Vector-Based Many-Objective Evolutionary Algorithms

A. DTLZ Test Problems

The DTLZ test suite was designed by Deb et al. [9] as a set of nine scalable test problems (DTLZ1-9). The number of decision variables (say {n} ) of {M} -objective DTLZ1-7 problems is specified as n=M+k-1 , where {k} is a parameter. The value of {k} is often specified as {k} \,\, = \,\, 5 in DTLZ1, {k} \,\, = \,\, 10 in DTLZ2-6, and {k} \,\, = \,\, 20 in DTLZ7. In DTLZ8-9, the number of decision variables is specified as {n} \,\, = \,\, 10{M} .

Let { \boldsymbol {y}}^{*} = (y^{*}_{1}, y^{*}_{2}, \ldots , y^{*}_{M}) be a Pareto optimal solution in the {M} -dimensional objective space of each test problem. The Pareto front of each of the first four test problems (DTLZ1-4) can be represented by the following formulations [9]:\begin{align}&{\mathrm{ DTLZ}}1:~\sum _{i=1}^{M}y_{i}^{_{*}} = 0.5~{\mathrm{ and}}~y_{i}^{_{*}} \ge 0~{\mathrm{ for}}~i = 1, 2, \ldots , M \\&{\mathrm{ DTLZ}}\hbox {2-4}:~\sum _{i=1}^{M}\left ({y_{i}^{_{*}} }\right )^{2} = 1~{\mathrm{ and}}~y_{i}^{_{*}} \ge 0~{\mathrm{ for}}~i = 1, 2, \ldots ,M. \notag \\ {}\end{align} View SourceRight-click on figure for MathML and additional features.

As shown in Table I, DTLZ1-4 have been frequently used for performance evaluation of many-objective algorithms. This is because their Pareto fronts are represented by the simple formulations in (2) and (3). DTLZ5-6 were originally proposed as many-objective test problems with degenerate Pareto fronts in [9]. However, their Pareto fronts are not degenerate when they have four or more objectives [10], [36], [37]. Constraint conditions were introduced to remove the nondegenerate parts of the Pareto fronts [36], [37]. The Pareto fronts of DTLZ7-9 cannot be represented by a simple form as in (2) or (3).

B. WFG Test Problems

The WFG test suite was proposed by Huband et al. [10] as a set of nine scalable test problems (WFG1-9). An {M} -objective WFG problem has {k} position-related variables and {l} distance-related variables. Thus the total number of decision variables is n = k + l . The suggested values of {k} and {l} in [10] were k = 4 and l = 20 for two-objective problems and k = 2(M - 1) and l = 20 for problems with more than two objectives. However, different specifications have been used in the literature. For example, they were specified as k = M - 1 and l = 24 - (M - 1) in [29]. These specifications [29] are used in this paper. Since {l} should be an even number in WFG2 and WFG3, the specification of {l} is slightly changed as {l} \,\, = \,\, 16 for {M} \,\, = \,\, 8 and {l} \,\, = \,\, 14 for M = 10 only in WFG2 and WFG3 in this paper.

WFG1 has a complicated Pareto front, which cannot be represented by a simple form as in (2) or (3). The Pareto front of WFG2 is disconnected. WFG3 was originally designed as a many-objective test problem with a degenerate Pareto front. However, its Pareto front is not degenerate when it has three or more objectives as recently pointed out in [38]. All the other test problems (i.e., WFG4-9) have the following Pareto front:\begin{align} {\mathrm{ WFG}}\hbox {4-9}:\sum _{i=1}^{M}\left ({\frac {y_{i}^{_{*}}}{2i} }\right )^{2} = 1~{\mathrm{ and}}~y_{i}^{_{*}} \ge 0~{\mathrm{ for}}~i = 1, 2, \ldots , M.\notag \\ {}\end{align} View SourceRight-click on figure for MathML and additional features.

One important feature of the Pareto fronts of WFG4-9 in (4) is that the domain of each objective has a different magnitude (e.g., 0\le y_{1}^{_{*}} \le 2 and 0\le y_{10}^{_{*}} \le 20 ). This explains why most of the recently proposed many-objective algorithms [27]–​[30], [32] in Table I have normalization mechanisms of the objective space. Since MOEA/D [25] has no normalization mechanism, good results have not been reported for many-objective WFG problems in the literature.

C. Common Feature of DTLZ1-4 and WFG4-9

By normalizing the range of the Pareto front for each objective into the unit interval [0, 1], the Pareto fronts of DTLZ1-4 and WFG4-9 can be rewritten as follows:\begin{align}&{\mathrm{ DTLZ}}1:~\sum _{i=1}^{M}y_{i}^{_{*}} = 1~{\mathrm{ and}}~y_{i}^{_{*}} \ge 0~{\mathrm{ for}}~i=1,2,\ldots , M\qquad \\&{\mathrm{ DTLZ}}\hbox {2-4}~{\mathrm{ and~WFG}}\hbox {4-9}\notag \\&\quad \qquad \sum _{i=1}^{M}\left ({y_{i}^{_{*}} }\right )^{2} = 1~{\mathrm{ and}}~y_{i}^{_{*}} \ge 0~{\mathrm{ for}}~i = 1, 2, \ldots ,M. \end{align} View SourceRight-click on figure for MathML and additional features.

These formulations show that DTLZ1-4 and WFG4-9 are similar test problems with respect to the shape of their Pareto fronts. This means that most test problems used for evaluating many-objective algorithms in [27]–​[32] in Table I are similar with respect to the shape of their Pareto fronts whereas they are different with respect to other aspects such as the curvature property of the Pareto fronts (e.g., linear and concave) and the type of the objective functions (e.g., multimodal and deceptive). In this paper, we explain our concern that the development of weight vector-based many-objective evolutionary algorithms seems to be overspecialized for the above-mentioned similarity of the Pareto fronts of DTLZ1-4 and WFG4-9.

SECTION III.

Weight Vector-Based Many-Objective Algorithms

In MOEA/D [25], a multiobjective problem is decomposed into single-objective problems, each of which is generated by a scalarizing function with a different weight vector. Thus the number of single-objective problems is the same as the number of weight vectors. Since a single best solution is stored for each single-objective problem, the population size is also the same as the number of weight vectors. MOEA/D can be viewed as an improved version of a cellular multiobjective genetic algorithm [39]. MOEA/D is also similar to multiobjective genetic local search (MOGLS) [40]–​[43]: both of them optimize scalarizing functions. Whereas weight vectors are systematically generated and fixed in MOEA/D, they are randomly updated in each generation in MOGLS.

A. Weight Vector Specification

In the original MOEA/D [25], all weight vectors { \boldsymbol {w}} = (w_{1}, w_{2}, \ldots , w_{M}) satisfying the following formulations are generated:\begin{align}&\sum _{i=1}^{M}w_{i} = 1~{\mathrm{ and}}~w_{i} \ge 0~{\mathrm{ for}}~i = 1, 2, \ldots , M \\&w_{i} \in \left \{{0,\frac {1}{H} ,\frac {2}{H} ,\ldots , \frac {H}{H} }\right \}~{\mathrm{ for}}~i = 1, 2, \ldots , M \end{align} View SourceRight-click on figure for MathML and additional features. where {H} is a positive integer. The total number of the weight vectors (say {N} , which is the same as the population size) is calculated as N=C_{M-1}^{H+M-1} (i.e., N = _{H+M-1}C_{M-1} [25]).

Weight vectors in (7) and (8) are on the hyper-plane specified by (7) in an {M} -dimensional space. For example, all weight vectors for {M} = 3 are on the triangular shape plane specified by w_{1} +w_{2} +w_{3} =1 and w_{i} \ge 0 for {i} = 1, 2, 3. When {H} \,\, < \,\, {M} , at least one element is zero (i.e., \textit{w}_{i} = 0) in all weight vectors satisfying (7) and (8). That is, no weight vector is inside the hyper-plane specified by (7). However, the use of a large value for {H} satisfying H\ge M leads to an impractically large number of weight vectors for many-objective problems. For example, if we specify {H} as H=10 for a ten-objective problem, 92 378 weight vectors are generated. In this case, only a single weight vector (0.1, 0.1, \ldots , 0.1 ) is inside the hyper-plane. All the other weight vectors are on its boundary.

For handling this difficulty, a two-layered approach was proposed in NSGA-III [30] and used in recently proposed weight vector-based many-objective algorithms. In the two-layered approach, two sets of weight vectors { \boldsymbol {w}} = (w_{1}, w_{2},\ldots , w_{M}) are generated using (7) and (8). One set of weight vectors is generated from an integer {H} _{1} and used with no modification as the boundary layer weight vectors. The other set of weight vectors is generated from another integer {H} _{2} and used as the inside layer weight vectors after the following modification:\begin{equation} w_{i} =\left ({w_{i} +1/M}\right )/2~{\mathrm{ for}}~i = 1, 2, \ldots , M. \end{equation} View SourceRight-click on figure for MathML and additional features.

The two-layered approach is used for many-objective problems with eight or more objectives in this paper. It should be noted that all weight vectors generated by the two-layered approach are on the same hyper-plane specified by (7), which is the same as all weight vectors in the original MOEA/D.

Another frequently used modification is the use of a small number 10^{-6} when w_{i} = 0 in the weighted Tchebycheff function. We use this modification in our computational experiments.

B. Basic Idea of MOEA/D

The basic idea of MOEA/D (and its variants) is to find a set of well-distributed nondominated solutions along the Pareto front using the systematically generated weight vectors as shown in Fig. 6. The realization of this idea is often explained by two distances in Fig. 7: 1) {d} _{1} and 2) {d} _{2} .

Fig. 6. - Basic idea of MOEA/D and weight vector-based algorithms.
Fig. 6. - Basic idea of MOEA/D and weight vector-based algorithms.
Fig. 6.

Basic idea of MOEA/D and weight vector-based algorithms.

Fig. 7. - Search for a solution using the weight vector 
$\boldsymbol {w}$
 in MOEA/D.
Fig. 7. - Search for a solution using the weight vector 
$\boldsymbol {w}$
 in MOEA/D.
Fig. 7.

Search for a solution using the weight vector \boldsymbol {w} in MOEA/D.

In Fig. 7, {d} _{1} is the distance from the ideal point \boldsymbol {z}^{I} to the solution ({f} _{1} (\boldsymbol {x} ), {f} _{2} (\boldsymbol {x} )) along the reference line specified by a weight vector, and {d} _{2} is the distance from the reference line to the solution ({f} _{1} (\boldsymbol {x} ), {f} _{2} (\boldsymbol {x} )). These distances should be minimized for each reference line (i.e., in each single-objective problem).

It is also important to assign a single solution to each reference line. Weight vector-based algorithms [27]–​[32] have their own mechanisms to minimize {d} _{1} and {d} _{2} , and to assign a single solution to each reference line. While some algorithms [30]–​[32] use Pareto dominance for minimizing {d} _{1} , the basic idea is the same: the search for well-distributed nondominated solutions along the Pareto front using a set of reference lines.

The ideal point \boldsymbol {z}^{I} is defined by the best value of each objective as shown in Fig. 7. In general, the ideal point \boldsymbol {z}^{I} is unknown. So, its approximation is usually used as a reference point \boldsymbol {z} ^{*} in MOEA/D and other weight vector-based algorithms as explained in the next section.

C. Scalarizing Functions

Three scalarizing functions were examined in the original MOEA/D [25]: 1) the weighted sum; 2) the weighted Tchebycheff function; and 3) the penalty-based boundary intersection (PBI) function. For the minimization problem in (1), the weighted sum with a weight vector { \boldsymbol {w}} = (w_{1}, w_{2}, \ldots , w_{M}) is written as \begin{equation} {\mathrm{ Minimize}}~f^{\mathrm{ WS}} ({ \boldsymbol {x}}|{ \boldsymbol {w}})=w_{1} f_{1} ({ \boldsymbol {x}})+\cdots +w_{M} f_{M} ({ \boldsymbol {x}}). \end{equation} View SourceRight-click on figure for MathML and additional features.

The weighted Tchebycheff function is written using a reference point { \boldsymbol {z}}^{_{*}} =(z_{1}^{_{*}} , z_{2}^{_{*}} ,\ldots , z_{M}^{_{*}} ) as \begin{equation} {\mathrm{ Minimize}}~f^{\mathrm{ Tch}} \left ({{ \boldsymbol {x}}|{ \boldsymbol {w}}, { \boldsymbol {z}}^{_{*}} }\right )=\max \limits _{i=1,2,\ldots ,M} \left \{{ w_{i} \cdot \left |{z_{i}^{_{*}} -f_{i} ({ \boldsymbol {x}})}\right |}\right \}. \qquad \end{equation} View SourceRight-click on figure for MathML and additional features.

Each element z_{i}^{_{*}} of the reference point { \boldsymbol {z}}^{_{*}} is specified by the minimum (i.e., best) value of each objective f_{i} ({ \boldsymbol {x}}) among the examined solutions during the execution of MOEA/D.

Using a penalty parameter \theta (\theta ={\mathrm{ 5}} in [25]) and the reference point { \boldsymbol {z}}^{_{*}} , the PBI function is written as \begin{equation} {\mathrm{ Minimize}}~f^{\mathrm{ PBI}} \left ({{ \boldsymbol {x}}|{ \boldsymbol {w}}, { \boldsymbol {z}}^{_{*}} }\right )=d_{1} +\theta \, d_{2} \end{equation} View SourceRight-click on figure for MathML and additional features. where {d} _{1} and {d} _{2} are defined as follows (see Fig. 7):\begin{align} d_{1}=&\left |{\left ({\,{ \boldsymbol {f}}({ \boldsymbol {x}})-{ \boldsymbol {z}}^{_{*}} }\right )^{T} { \boldsymbol {w}}}\right | \big / \left \|{ { \boldsymbol {w}}}\right \| \\ d_{2}=&\left \|{ \,{ \boldsymbol {f}}({ \boldsymbol {x}})-{ \boldsymbol {z}}^{_{*}} -d_{1} \frac {{ \boldsymbol {w}}}{||{ \boldsymbol {w}}||} }\right \|. \end{align} View SourceRight-click on figure for MathML and additional features.

D. Neighborhood Structure

In the original MOEA/D [25], each weight vector has its neighbors. A prespecified number of similar weight vectors are defined as neighbors for each weight vector. A weight vector itself is included in its own neighbors. When a solution is to be generated for a weight vector, parents are selected from its neighbors. The generated new solution is compared with the solution of each neighbor. If the new solution is better, the current solution is replaced. The comparison for replacement is performed against all solutions of the neighbors.

This replacement strategy has two potential difficulties. One is explained by the following case. A new solution, which is generated far from the corresponding reference line, is very good for a different reference line while its evaluation is poor for the corresponding reference line. This case is explained in Fig. 8. Let us assume in Fig. 8 that the set of the seven open circles is a current population. We also assume that solution A is generated for a reference line {l} _{2} with two neighbors {l} _{1} and {l} _{3} . The generated solution A, which is very good for a reference line {l} _{7} , is not good for any of {l} _{1} , {l} _{2} , and {l} _{3} . Thus no solution is replaced with A. While such a situation is not likely to happen very often, its appropriate handling may improve the efficiency of MOEA/D. This difficulty can be partially remedied by using a large number of neighbors. However, this remedy may worsen another potential difficulty explained by the following case: many neighboring solutions are replaced with a single good solution. This leads to the decrease in the diversity of solutions. For example, if solution B is generated for the reference line {l} _{2} in Fig. 8, all the current solutions of {l} _{1} , {l} _{2} , and {l} _{3} will be replaced with solution B.

Fig. 8. - Illustration of reference lines and generated solutions A and B.
Fig. 8. - Illustration of reference lines and generated solutions A and B.
Fig. 8.

Illustration of reference lines and generated solutions A and B.

E. Performance Improvement Mechanisms

These two difficulties can be remedied by the following replacement policy. A new solution is compared with its similar solutions, and only a single solution is replaced with the new solution. Instead of using the prespecified neighbors, a set of similar solutions is selected for the new solution. First the nearest reference line to the new solution is identified in the objective space (e.g., line {l} _{2} for solution B in Fig. 8). Next all solutions with the same nearest reference line as the new solution are selected as its similar solutions. Then the new solution is compared with each similar solution. Only a single solution can be replaced with the new solution. If the new solution is not better than any similar solutions, no solution is replaced. This policy is implemented in [27]–​[32] using different mechanisms for similar solution selection, solution comparison, and solution replacement.

In NSGA-III [30], reference points are specified in the normalized objective space in a similar manner to the weight vector specification in MOEA/D. Using the reference points and the origin of the normalized objective space, reference lines are generated. Each solution is assigned to its nearest reference line. Solution comparison in NSGA-III is performed using nondominated sorting, the number of solutions assigned to each reference line, and the distance from each solution to its nearest reference line. Solutions are updated by a (\mu +\mu )ES-style model. When multiple solutions with the best nondominated sorting rank are assigned to the same reference line, the best solution with the shortest distance to the reference line is selected from them as a member in the next generation. The other solutions are randomly ordered. After the generation update, most reference lines have a single solution. However, it is possible that some reference lines have multiple solutions. It is also possible that other reference lines have no solutions.

In \theta -DEA [29], each solution is assigned to its nearest reference line in the same manner as NSGA-III. The PBI function is used for the ranking of solutions assigned to the same reference line. Solutions are updated by a (\mu +\mu ) ES-style model based on the rank of each solution. Some reference lines may have multiple solutions, and others may have no solutions.

In MOEA/DD [31], a (\mu +1 )ES-style model is used as in the original MOEA/D. Each solution is assigned to the nearest weight vector as in NSGA-III and \theta -DEA. Each solution is evaluated by its nondominated sorting rank, its PBI value, and the number of solutions assigned to the same weight vector. When all solutions in the current population are nondominated, one solution with the largest PBI value is deleted from the most crowded weight vector with the largest number of solutions. When some solutions are dominated, one solution to be deleted is selected from the worst rank solutions using the number of solutions assigned to each weight vector and the PBI value of each solution. However, when a weight vector has a single solution, the solution is not deleted for diversity maintenance even if it has the worst nondominated sorting rank.

As we have already explained, weight vector-based algorithms in [27]–​[30] and [32] have normalization mechanisms. The two-layered approach in NSGA-III [30] are used in [27]–​[32]. In addition to these two common features, each algorithm has its own mechanisms for efficiently realizing the search for a set of well-distributed solutions along the Pareto front of a many-objective problem.

F. Relation Between Test Problems and Algorithms

As shown in Section II, the DTLZ1-4 and WFG4-9 test problems have the following Pareto front in the normalized {M} -dimensional objective space:\begin{align} {\mathrm{ Pareto~Front}}:~\sum _{i=1}^{M}\left ({y_{i}^{_{*}} }\right )^{k} = 1~{\mathrm{ and}}~y_{i}^{_{*}} \ge 0~{\mathrm{ for}}~i = 1, 2, \ldots , M \notag \\ {}\end{align} View SourceRight-click on figure for MathML and additional features. where k = 1 (DTLZ1) and k = 2 (DTLZ2-4 and WFG4-9).

Weight vectors in the weight vector-based algorithms [27]–​[32] are generated in the {M} -dimensional weight space as \begin{align} {\mathrm{ Weight~Vectors}}:\sum _{i=1}^{M}w_{i} = 1~{\mathrm{ and}}~w_{i} \ge 0~{\mathrm{ for}}~i = 1, 2, \ldots , M. \notag \\ {}\end{align} View SourceRight-click on figure for MathML and additional features.

The shape of the Pareto front of each test problem in (15) is the same as or similar to the shape of the hyper-plane in (16) on which the weight vectors are generated. This explains why the search for a single best solution for each weight vector leads to a set of well-distributed solutions over the entire Pareto front. From the similarity between the shape of the Pareto front in (15) and the shape of the distribution of the weight vectors in (16), the following questions may arise.

  1. How genral is the high performance of weight vector-based algorithms on the DTLZ1-4 and WFG4-9 problems?

  2. How general are the triangular shape Pareto fronts of the DTLZ1-4 and WFG4-9 test problems?

In the next section, we discuss the first question through computational experiments. The second question is briefly discussed in Section V as a future research topic.

SECTION IV.

Computational Experiments

In this section, first we explain our test problems generated by slightly changing the DTLZ and WFG formulations. Then we demonstrate how high performance of weight vector-based algorithms on the DTLZ and WFG problems is deteriorated by the slight change in the test problem formulations.

A. Our Test Problems: DTLZ^{-1} and WFG^{-1}

As we have already explained, the DTLZ and WFG test problems have the following form:\begin{equation} {\mathrm{ Minimize}}~f_{1} ({ \boldsymbol {x}}), \ldots , f_{M} ({ \boldsymbol {x}})~{\mathrm{ subject~to}}~{ \boldsymbol {x}}\in { \boldsymbol {X}}. \end{equation} View SourceRight-click on figure for MathML and additional features.

Our idea is to generate a slightly different test problem from each of the DTLZ and WFG problems. More specifically, we change their general form from (17) to \begin{equation} {\mathrm{ Maximize}}~f_{1} ({ \boldsymbol {x}}), \ldots , f_{M} ({ \boldsymbol {x}})~{\mathrm{ subject~to}}~{ \boldsymbol {x}}\in { \boldsymbol {X}}. \end{equation} View SourceRight-click on figure for MathML and additional features.

We use exactly the same objective functions and the same constraint conditions except for changing from “Minimize” in (17) to “Maximize” in (18). The generated problems in (18) are referred to as the Max-DTLZ and Max-WFG problems. Those problems are handled as the following minimization problems:\begin{equation} {\mathrm{ Minimize}} \,\, -f_{1} ({ \boldsymbol {x}}), \ldots , -f_{M} ({ \boldsymbol {x}})~{\mathrm{ subject~to}}~{ \boldsymbol {x}}\in { \boldsymbol {X}}. \end{equation} View SourceRight-click on figure for MathML and additional features.

All objectives in DTLZ and WFG are multiplied by (–1) in our test problems in (19). That is, a negative sign is added to all objectives in DTLZ and WFG. In this paper, the minus versions in (19) of DTLZ and WFG are referred to as DTLZ^{-1} and WFG^{-1} (Minus-DTLZ and Minus-WFG), respectively. The minimization formulation in (19) with a minus sign for all objectives is the same as the maximization form in (18). In our computational experiments, we use the minimization form to handle all test problems as minimization problems (i.e., to apply each algorithm with no modification to DTLZ, WFG, DTLZ^{-1} , and WFG^{-1} in exactly the same manner).

In Fig. 9, we show the Pareto fronts of the original three-objective DTLZ2 problem and its maximization version in (18): Max-DTLZ2. The Pareto fronts of the two test problems have the same shape with different size. In Fig. 10, we show the Pareto fronts of the minus versions of the three-objective DTLZ1 and DTLZ2 (i.e., DTLZ1^{-1} and DTLZ2^{-1} ). As shown in Fig. 10(a), DTLZ1^{-1} has a rotated triangular shape Pareto front. The Pareto front of DTLZ2^{-1} in Fig. 10(b) has a rotated shape of the Pareto front of the maximization version of DTLZ2 in Fig. 9(b). In this section, we examine the performance of weight vector-based algorithms on the DTLZ and WFG test problems and their minus versions in (19).

Fig. 9. - Pareto fronts of three-objective DTLZ2 and Max-DTLZ2 problems. (a) Original three-objective DTLZ2. (b) Maximization version of DTLZ2.
Fig. 9. - Pareto fronts of three-objective DTLZ2 and Max-DTLZ2 problems. (a) Original three-objective DTLZ2. (b) Maximization version of DTLZ2.
Fig. 9.

Pareto fronts of three-objective DTLZ2 and Max-DTLZ2 problems. (a) Original three-objective DTLZ2. (b) Maximization version of DTLZ2.

Fig. 10. - Pareto fronts of the minus versions of DTLZ1 and DTLZ2. (a) Three-objective DTLZ
$1^{-1}$
. (b) Three-objective DTLZ
$2^{-1}$
.
Fig. 10. - Pareto fronts of the minus versions of DTLZ1 and DTLZ2. (a) Three-objective DTLZ
$1^{-1}$
. (b) Three-objective DTLZ
$2^{-1}$
.
Fig. 10.

Pareto fronts of the minus versions of DTLZ1 and DTLZ2. (a) Three-objective DTLZ1^{-1} . (b) Three-objective DTLZ2^{-1} .

As shown in Figs. 9 and 10 for DTLZ1 and DTLZ2, the minus versions of DTLZ have much larger Pareto fronts than the original DTLZ problems. When DTLZ has a concave Pareto front, its minus version has a convex Pareto front. Moreover, the difficulty of optimization is not the same. For example, the Pareto optimal solutions of DTLZ2 are obtained when all of its distance variables are 0.5. However, the Pareto optimal solutions of DTLZ2^{-1} are obtained when all of its distance variables are 0 or 1 (i.e., at the boundaries of the feasible region of each distance variable x_{i} with the constraint condition 0\le x_{i} \le 1 ). These discussions clearly show that our idea is not the best way for inverting the Pareto front without changing the other properties of the original DTLZ and WFG test problems. For this purpose, modification of only the shape function may be the best way. In this paper, we use the above-mentioned simple idea for inverting the Pareto front since it is simple and intuitively understandable.

A test problem called the inverted DTLZ1 was formulated in [44] by applying the following transformation to each objective f_{i} ({ \boldsymbol {x}}) of DTLZ1 (i = 1, 2, \ldots , M ):\begin{equation} f_{i} ({ \boldsymbol {x}})\leftarrow 0.5\, (1+g({ \boldsymbol {x}}))-f_{i} ({ \boldsymbol {x}}) \end{equation} View SourceRight-click on figure for MathML and additional features. where {g} (\boldsymbol {x} ) is a function in the original DTLZ1 formulation [9]. The Pareto front of the inverted DTLZ1 [44] has a rotated shape of the Pareto front of DTLZ1 [9]. The size of their Pareto fronts is the same since the inverted DTLZ1 was formulated by rotating DTLZ1. It should be noted that our DTLZ1^{-1} has a much larger Pareto front than DTLZ1 and its rotated version as shown in Fig. 10(a).

B. Examined Algorithms and Parameter Specifications

Our computational experiments on DTLZ1-4 and WFG1-9 with 3, 5, 8, and 10 objectives are performed under the same settings as in [29]. We examine the performance of \theta -DEA [29], NSGA-III [30], and MOEA/DD [31]. We also use NSGA-II [7] and four versions of MOEA/D [25]: 1) MOEA/D-WS with the weighted sum; 2) MOEA/D-Tch with the Tchebycheff function; 3) MOEA/D-PBI with the PBI function; and 4) MOEA/D-IPBI with the inverted PBI (IPBI) function [45].

The IPBI function is defined by the distance from the nadir point \boldsymbol {z}^{N} , which is a vector consisting of the worst value of each objective over the true Pareto front. For minimization problems, the IPBI function [45] is defined as \begin{equation} {\mathrm{ Maximize}}~f^{\mathrm{ IPBI}} \left ({{ \boldsymbol {x}}|{ \boldsymbol {w}}, { \boldsymbol {z}}^{N} }\right )=d_{1} -\theta \, d_{2} \end{equation} View SourceRight-click on figure for MathML and additional features. where \theta is a penalty parameter, and {d} _{1} and {d} _{2} are defined as follows:\begin{align} d_{1}=&\left |{\left ({{ \boldsymbol {z}}^{N} -{ \boldsymbol {f}}({ \boldsymbol {x}})}\right )^{T} { \boldsymbol {w}}}\right | / \left \|{ { \boldsymbol {w}}}\right \| \\ d_{2}=&\left \|{ { \boldsymbol {z}}^{_{N}} -{ \boldsymbol {f}}({ \boldsymbol {x}})-d_{1} \frac {{ \boldsymbol {w}}}{||{ \boldsymbol {w}}||} \, }\right \|. \end{align} View SourceRight-click on figure for MathML and additional features.

In this paper, we use the same implementation of MOEA/D-IPBI as in [45]: the penalty parameter \theta is specified as \theta = 0.1 , and the nadir point \boldsymbol {z}^{N} is approximated by the worst value of each objective in the current population.

Multiobjective search is performed in MOEA/D-IPBI by pushing each solution in the direction from the nadir point to the Pareto front by maximizing {d} _{1} and minimizing {d} _{2} . This search mechanism contrasts to MOEA/D-PBI, MOEA/D-Tch and the other weight vector-based algorithms in [27]–​[32] where each solution is pulled toward the ideal point.

Exactly the same set of weight vectors is used in all algorithms for each test problem. Table II shows the number of weight vectors. The two-layered approach is used for test problems with eight and ten objectives. The population size is the same as the number of weight vectors except for NSGA-II, NSGA-III, and \theta -DEA where the population size is specified as multiples of four (see Table II). The weight value 0 is replaced with 10^{-6} in MOEA/D-Tch.

TABLE II Number of Weight Vectors and the Population Size for Each Test Problem (the Same Specifications as in [29])
Table II- 
Number of Weight Vectors and the Population Size for Each Test Problem (the Same Specifications as in [29])
Table II- 
Number of Weight Vectors and the Population Size for Each Test Problem (the Same Specifications as in [29])

As a termination condition for each test problem, we use the same prespecified total number of generations as in [29]. The penalty parameter \theta in MOEA/D-PBI, MOEA/DD and \theta -DEA is specified as \theta = 5 (\theta = 0.1 in IPBI). The neighborhood size in MOEA/D is 20. The polynomial mutation with the distribution index 20 is used with the mutation probability 1/{n} ({n} is the string length). The simulated binary crossover with the distribution index 20 (30 in NSGA-III, \theta -DEA and MOEA/DD) is used with the crossover probability 1.0. Our parameter specifications are the same as in [29].

We use our own implementations of the four versions of MOEA/D because we have already examined them on various test problems. This is also because we have not observed any clear inconsistency between our results and the reported results in [29]. With respect to NSGA-II, \theta -DEA, NSGA-III, and MOEA/DD, we use available codes through the Internet: NSGA-II from jMetal [46], \theta -DEA, and NSGA-III from [47] by Yuan et al. [29] of the \theta -DEA paper, and MOEA/DD from [48] by Li et al. [31] of the MOEA/DD paper.

C. Performance Measures

As a performance measure, we use the hypervolume in the same manner as [29] for the original DTLZ and WFG problems (the setting of the reference point for hypervolume calculation is the same as [29]). First, the objective space is normalized using the ideal point \boldsymbol {z}^{I} and the nadir point \boldsymbol {z}^{N} so that they are normalized as { \boldsymbol {z}}^{I} = (0, 0, \ldots , 0) and { \boldsymbol {z}}^{N} = (1, 1, \ldots , 1) . Then the hypervolume is calculated by specifying the reference point as (1.1, 1.1, \ldots , 1.1) . The same setting for hypervolume calculation is used for DTLZ^{-1} and WFG^{-1} . We also use a different reference point (2, 2, \ldots , 2 ) for DTLZ^{-1} and WFG^{-1} to examine whether well distributed solutions are obtained around the boundaries of their Pareto fronts. We use a fast calculation method of the exact hypervolume value proposed in [49] for all test problems with 3, 5, 8, and 10 objectives.

We also calculate the IGD indicator in the normalized objective space for all test problems. The Euclidean distance is used in the normalized objective space of each test problem. A set of reference points for the IGD calculation is generated for each test problem as follows. For DTLZ1-4, DTLZ1-4^{-1} , WFG4-9, and WFG3-9^{-1} , we randomly generate 100 000 reference points on the true Pareto front using the uniform distribution. For the other test problems (i.e., WFG1-3 and WFG1-2^{-1} ), we use all nondominated solutions among all of the obtained solutions in Tables III and IV since the specification of the uniform distribution over the true Pareto front is difficult for these test problems.

TABLE III Average Hypervolume Values for the Reference Point ( 1.1, 1.1, \ldots , 1.1 ) Over 101 Runs on the Original Test Problems. The Best Average Result for Each Test Problem is Highlighted by Bold and Underlined. The Worst Four Average Results for Each Test Problem are Shaded
Table III- 
Average Hypervolume Values for the Reference Point (
$1.1, 1.1, \ldots , 1.1$
) Over 101 Runs on the Original Test Problems. The Best Average Result for Each Test Problem is Highlighted by Bold and Underlined. The Worst Four Average Results for Each Test Problem are Shaded
Table III- 
Average Hypervolume Values for the Reference Point (
$1.1, 1.1, \ldots , 1.1$
) Over 101 Runs on the Original Test Problems. The Best Average Result for Each Test Problem is Highlighted by Bold and Underlined. The Worst Four Average Results for Each Test Problem are Shaded
TABLE IV Average Hypervolume Values Over 101 Runs on Our Minus Test Problems for the Reference Point ( 1.1, 1.1, \ldots ,~1.1 ). The Best Average Result for Each Test Problem is Highlighted by Bold and Underlined. The Worst Four Average Results for Each Test Problem are Shaded
Table IV- 
Average Hypervolume Values Over 101 Runs on Our Minus Test Problems for the Reference Point (
$1.1, 1.1, \ldots ,~1.1$
). The Best Average Result for Each Test Problem is Highlighted by Bold and Underlined. The Worst Four Average Results for Each Test Problem are Shaded
Table IV- 
Average Hypervolume Values Over 101 Runs on Our Minus Test Problems for the Reference Point (
$1.1, 1.1, \ldots ,~1.1$
). The Best Average Result for Each Test Problem is Highlighted by Bold and Underlined. The Worst Four Average Results for Each Test Problem are Shaded

D. Experimental Results on Original Test Problems

The average hypervolume values over 101 runs on the original DTLZ1-4 and WFG1-9 problems are summarized in Table III. The best average result is highlighted by bold and underlined for each test problem. The worst four average results are shaded. For DTLZ1-4, the best average results are obtained by MOEA/DD for 12 out of the 16 problems (75%). For WFG4-9, the best average results are obtained by \theta -DEA for 23 out of the 24 problems (96%). In Table III, these two algorithms are the best. The best average result for each test problem in WFG1-3 is obtained by a different algorithm due to their different features as test problems.

The best average result is not obtained by MOEA/D or MOEA/DD for WFG4-9 due to the lack of an objective space normalization mechanism. The performance of MOEA/DD on WFG4-9 can be improved by a normalization mechanism. Just for comparison, we perform computational experiments on the normalized WFG4-9 test problems after changing all objectives as {f_{i}}({ \boldsymbol {x}}) = f_{i}({ \boldsymbol {x}})/2i, i = 1, 2, \ldots , M . For the normalized WFG4-9 test problems, the best average results are obtained by MOEA/DD for 14 out of the 24 test problems (58%). This observation together with the results on DTLZ1-4 in Table III suggests that MOEA/DD with a normalized mechanism may be the best algorithm for DTLZ1-4 and WFG4-9.

E. Experimental Results on Modified Test Problems

Table IV shows the average hypervolume values for the reference point (1.1, 1.1, \ldots , 1.1 ) over 101 runs on our minus test problems: DTLZ^{-1} and WFG^{-1} . From the comparison between Tables III and IV, we can see that totally different results are obtained with respect to the performance comparison among the eight algorithms. For example, the best average results are obtained by \theta -DEA for 27 out of the 52 test problems (52%) in Table III, but only for six test problems (12%) in Table IV. In Table IV, NSGA-III looks the best. The best average results are obtained by NSGA-III for 19 test problems (37%). The relative performance of MOEA/D-WS is much better in Table IV than in Table III. An interesting observation is that the performance of NSGA-II is not always bad for many-objective test problems in Table IV.

The difference in the test problems between Tables III and IV is only the multiplication of (−1) to each objective. Except for this change, the test problems are the same between the two tables. However, totally different results are obtained. Especially, the performance of the best algorithms in Table III (i.e., \theta -DEA and MOEA/DD) is deteriorated in Table IV.

In Table V, we show experimental results evaluated by the hypervolume with the reference point (2, 2, \ldots , 2 ). Totally different results are obtained between Tables III (on DTLZ and WFG) and V (on DTLZ^{-1} and WFG^{-1} ) with respect to algorithm comparison. That is, good results are obtained by the last four algorithms in Table V for DTLZ^{-1} and WFG^{-1} while good results are obtained by the first four algorithms in Table III for DTLZ and WFG. An interesting observation is that better results in Table V are obtained from NSGA-II than NSGA-III for WFG4-9^{-1} while better results are obtained from NSGA-III than NSGA-II for WFG4-9 in Table III.

TABLE V Average Hypervolume Values Over 101 Runs on Our Minus Test Problems for the Reference Point ( 2, 2, \ldots , 2 ). The Best Average Result for Each Test Problem is Highlighted by Bold and Underlined. The Worst Four Average Results for Each Test Problem are Shaded
Table V- 
Average Hypervolume Values Over 101 Runs on Our Minus Test Problems for the Reference Point (
$2, 2, \ldots , 2$
). The Best Average Result for Each Test Problem is Highlighted by Bold and Underlined. The Worst Four Average Results for Each Test Problem are Shaded
Table V- 
Average Hypervolume Values Over 101 Runs on Our Minus Test Problems for the Reference Point (
$2, 2, \ldots , 2$
). The Best Average Result for Each Test Problem is Highlighted by Bold and Underlined. The Worst Four Average Results for Each Test Problem are Shaded

Performance evaluation results by the IGD indicator are shown in Table VI for DTLZ and WFG and Table VII for DTLZ^{-1} and WFG^{-1} . We can obtain similar observations from the IGD-based comparison results and the hypervolume-based comparison results. For example, the best results for almost all DTLZ and WFG problems are obtained from the first four algorithms in Tables III and VI whereas the best results for almost all DTLZ^{-1} and WFG^{-1} problems are obtained from the last four algorithms in Tables V and VII. We can also observe very bad results (i.e., very large average IGD values) of NSGA-II on DTLZ1-4 in Table VI.

TABLE VI Average IGD Values Over 101 Runs on the Original Test Problems. The Best Average Result for Each Test Problem is Highlighted by Bold and Underlined. The Worst Four Average Results for Each Test Problem are Shaded
Table VI- 
Average IGD Values Over 101 Runs on the Original Test Problems. The Best Average Result for Each Test Problem is Highlighted by Bold and Underlined. The Worst Four Average Results for Each Test Problem are Shaded
Table VI- 
Average IGD Values Over 101 Runs on the Original Test Problems. The Best Average Result for Each Test Problem is Highlighted by Bold and Underlined. The Worst Four Average Results for Each Test Problem are Shaded
TABLE VII Average IGD Values Over 101 Runs on Our Minus Test Problems. The Best Average Result for Each Test Problem is Highlighted by Bold and Underlined. The Worst Four Average Results for Each Test Problem are Shaded
Table VII- 
Average IGD Values Over 101 Runs on Our Minus Test Problems. The Best Average Result for Each Test Problem is Highlighted by Bold and Underlined. The Worst Four Average Results for Each Test Problem are Shaded
Table VII- 
Average IGD Values Over 101 Runs on Our Minus Test Problems. The Best Average Result for Each Test Problem is Highlighted by Bold and Underlined. The Worst Four Average Results for Each Test Problem are Shaded

SECTION V.

Discussion and Future Research Topics

A. Discussion on Experimental Results

For further discussing the experimental results in Tables III–​VII, we show a single run result of each algorithm on the three-objective DTLZ1^{-1} problem in Fig. 11. Among 101 runs, we select a single run with the median hypervolume value in Table IV to show a typical result. Fig. 11 shows all solutions in the final generation of the selected run. In Fig. 11, three values of \theta are examined for MOEA/D-IPBI (\theta = 0.1 was used in Tables III–​VII based on the reported results in [45]).

Fig. 11. - Experimental results of a single run of each algorithm on the three-objective DTLZ
$1^{-1}$
 problem. (a) NSGA-III. (b) 
$\theta $
-DEA. (c) MOEA/DD. (d) MOEA/D-PBI. (e) MOEA/D-Tch. (f) MOEA/D-WS. (g) MOEA/D-IPBI (
$\theta = 0.1$
). (h) NSGA-II. (i) MOEA/D-IPBI (
$\theta = 1$
). (j) MOEA/D-IPBI (
$\theta = 5$
).
Fig. 11. - Experimental results of a single run of each algorithm on the three-objective DTLZ
$1^{-1}$
 problem. (a) NSGA-III. (b) 
$\theta $
-DEA. (c) MOEA/DD. (d) MOEA/D-PBI. (e) MOEA/D-Tch. (f) MOEA/D-WS. (g) MOEA/D-IPBI (
$\theta = 0.1$
). (h) NSGA-II. (i) MOEA/D-IPBI (
$\theta = 1$
). (j) MOEA/D-IPBI (
$\theta = 5$
).
Fig. 11.

Experimental results of a single run of each algorithm on the three-objective DTLZ1^{-1} problem. (a) NSGA-III. (b) \theta -DEA. (c) MOEA/DD. (d) MOEA/D-PBI. (e) MOEA/D-Tch. (f) MOEA/D-WS. (g) MOEA/D-IPBI (\theta = 0.1 ). (h) NSGA-II. (i) MOEA/D-IPBI (\theta = 1 ). (j) MOEA/D-IPBI (\theta = 5 ).

Well-distributed solutions are obtained by MOEA/D-IPBI in Fig. 11(i) and (j). Such a good solution set is not obtained by the other algorithms. In Fig. 11(d), ten solutions are obtained inside the Pareto front together with many solutions on its boundary. This result is explained by the inconsistency between the shape of the Pareto front and the shape of the distribution of the weight vectors in Fig. 12. The shaded region is the projection of the Pareto front, which is the region of weight vectors intersecting with the Pareto front.

Fig. 12. - Relation between a set of weight vectors and the Pareto front.
Fig. 12. - Relation between a set of weight vectors and the Pareto front.
Fig. 12.

Relation between a set of weight vectors and the Pareto front.

Almost the same figure as Fig. 12 was used in [44] for explaining experimental results of NSGA-III on the inverted DTLZ1 problem. As shown in Fig. 12 (and in [44]), the weight vectors are uniformly distributed over the triangle whereas the shape of the Pareto front is a rotated triangle. We can see from Fig. 12 that the ten weight vectors are inside the projection of the Pareto front. Those weight vectors correspond to the ten inside solutions in Fig. 11(d). Since each weight vector in MOEA/D-PBI always has a single solution, many solutions are obtained on the boundary of the Pareto front in Fig. 11(d). Those boundary solutions are the best solutions for the outside weight vectors in Fig. 12. For the same reason, many solutions are obtained on the boundary in Fig. 11(e).

Multiobjective search in weight vector-based algorithms in [27]–​[32] as well as MOEA/D-PBI and MOEA/D-Tch can be viewed as pulling all solutions toward the ideal point using the weight vectors. The weight vectors in those algorithms are illustrated in Fig. 13(a). Figs. 12 and 13(a) show the same weight vectors. The ten inside solutions in Fig. 11(d) are obtained by the ten inside weight vectors in Fig. 12. Almost the same ten inside solutions are obtained in Fig. 11(a)–(d) since all algorithms in Fig. 11(a)–(d) have the same basic idea of multiobjective search: pulling all solutions toward the ideal point using the weight vectors in Fig. 13(a).

Fig. 13. - Weight vectors used for three-objective minimization. (a) Most weight vector-based algorithms. (b) Weighted sum and IPBI.
Fig. 13. - Weight vectors used for three-objective minimization. (a) Most weight vector-based algorithms. (b) Weighted sum and IPBI.
Fig. 13.

Weight vectors used for three-objective minimization. (a) Most weight vector-based algorithms. (b) Weighted sum and IPBI.

When IPBI is used, multiobjective search is performed by pushing all solutions from the nadir point to the Pareto front as shown in Fig. 13(b). In this case, the distribution of weight vectors (i.e., reference lines) is consistent with the shape of the Pareto front. This explains why well-distributed solutions are obtained by MOEA/D-IPBI in Fig. 11(i) and (j).

In NSGA-III, \theta -DEA, and MOEA/DD, each solution is assigned to its nearest reference line (whereas each reference line has its best solution in MOEA/D). If there is no solution close to a reference line, no solution is assigned. In Fig. 12, all weight vectors outside the shaded region have no solution. As a result, many solutions are not obtained on the boundary of the Pareto front by NSGA-III, \theta -DEA, and MOEA/DD. In these algorithms, a single reference line can have multiple solutions. This is the reason why more inside solutions are obtained in Fig. 11(a)–(c) than in Fig. 11(d). In NSGA-III, the second solution for each reference line is selected randomly from the assigned solutions with the best rank. As a result, solution distribution looks somewhat random in Fig. 11(a). Since the second solution for each reference line in \theta -DEA is the second best solution with respect to PBI, all solutions assigned to the same reference line are almost the same as shown in Fig. 11(b). Similar solution sets are obtained in Fig. 11(b) and (c) since \theta -DEA and MOEA/DD have similar mechanisms to choose the second best solution for each reference line.

In the same manner as Fig. 11, we show an experimental result of a single run of each algorithm on the three-objective DTLZ2^{-1} problem in Fig. 14. Good results are obtained by MOEA/D-WS and MOEA/D-IPBI in Fig. 14. One difference between Figs. 11 and 14 is the results by MOEA/D-WS and MOEA/D-IPBI (\theta =0.1 ). Since the Pareto front of DTLZ1^{-1} is a plane, well-distributed solutions are not obtained in Fig. 11 by MOEA/D-WS and MOEA/D-IPBI (\theta =0.1 ). However, good results are obtained by these algorithms in Fig. 14 since the Pareto front of DTLZ2^{-1} is convex. It should be noted that the weighted sum and the IPBI function with \theta =0.1 have similar contour lines [45].

Fig. 14. - Experimental results on the three-objective DTLZ
$2^{-}{}^{1}$
 test problem. (a) NSGA-III. (b) 
$\theta $
-DEA. (c) MOEA/DD. (d) MOEA/D-PBI. (e) MOEA/D-Tch. (f) MOEA/D-WS. (g) MOEA/D-IPBI (
$\theta = 0.1$
). (h) NSGA-II. (i) MOEA/D-IPBI (
$\theta = 1$
). (j) MOEA/D-IPBI (
$\theta = 5$
).
Fig. 14. - Experimental results on the three-objective DTLZ
$2^{-}{}^{1}$
 test problem. (a) NSGA-III. (b) 
$\theta $
-DEA. (c) MOEA/DD. (d) MOEA/D-PBI. (e) MOEA/D-Tch. (f) MOEA/D-WS. (g) MOEA/D-IPBI (
$\theta = 0.1$
). (h) NSGA-II. (i) MOEA/D-IPBI (
$\theta = 1$
). (j) MOEA/D-IPBI (
$\theta = 5$
).
Fig. 14.

Experimental results on the three-objective DTLZ2^{-}{}^{1} test problem. (a) NSGA-III. (b) \theta -DEA. (c) MOEA/DD. (d) MOEA/D-PBI. (e) MOEA/D-Tch. (f) MOEA/D-WS. (g) MOEA/D-IPBI (\theta = 0.1 ). (h) NSGA-II. (i) MOEA/D-IPBI (\theta = 1 ). (j) MOEA/D-IPBI (\theta = 5 ).

As in Fig. 11, the distribution of the weight vectors is not consistent with the shape of the Pareto front in Fig. 14 except for MOEA/D-WS and MOEA/D-IPBI. Thus, many solutions are obtained on the boundary of the Pareto front in Fig. 14(d) and (e). However, the distribution of solutions around the center of the Pareto front is similar between two groups. One is Fig. 14(b)–(d) with the inconsistent weight vector distribution, and the other is Fig. 14(f), (g), (i), and (j) with the consistent weight vector distribution. This can be explained by the convexity of the Pareto front of DTLZ2^{-1} . As shown in Fig. 6, when the Pareto front is convex, more solutions are obtained around its center by pulling solutions toward the ideal point. In this case, less solutions are obtained around the center by pushing solutions from the nadir point as shown in Fig. 14(f), (g), (i), and (j). As a result of these two effects (i.e., the inconsistency between the weight vector distribution and the Pareto front shape, and the Pareto front convexity), the distribution of solutions around the center of the Pareto front is similar between Fig. 14(b)–(d) and Fig. 14(f), (g), (i), and (j).

Thanks to random selection of the second solution for each reference point in NSGA-III, more solutions are obtained around the center of the Pareto front in Fig. 14(a) than the other algorithms in Fig. 14. This observation in Fig. 14(a) explains why good average results are obtained by NSGA-III in Table IV where the hypervolume is measured from the reference point (1.1, 1.1, \ldots , 1.1 ).

Hypervolume contributions of solutions around the center of the Pareto front are relatively large when the reference point is close to the Pareto front. By moving a reference point away from the Pareto front, their contributions become relatively small since contributions of solutions on the boundary of the Pareto front increase. This explains why the evaluation results of NSGA-III are not good in Table V with the reference point (2, 2, \ldots , 2 ). The difference in the evaluation results of NSGA-III between Tables IV and V suggests that many solutions around the center of the Pareto front are obtained by NSGA-III as shown in Figs. 11(a) and 14(a).

We also show an experimental result of a single run of each algorithm on the ten-objective DTLZ2^{-1} and WFG9^{-1} problems in Figs. 15 and 16, respectively. In Fig. 15, a wide variety of solutions are not obtained by the first four algorithms. This observation explains why those algorithms have low evaluation results on the ten-objective DTLZ2^{-1} problem in Table V with the reference point (2, 2, \ldots , 2 ) for hypervolume calculation whereas some of them have high evaluation results in Table IV with the reference point (1.1, 1.1, \ldots , 1.1 ). This is also the case in Fig. 16. Thanks to a large diversity of solutions in Figs. 15(h) and 16(h), NSGA-II has the best results on the ten-objective DTLZ2^{-1} and WFG9^{-1} problems in Table V.

Fig. 15. - Experimental results on the ten-objective DTLZ
$2^{-1}$
 test problem. (a) NSGA-III. (b) 
$\theta $
-DEA. (c) MOEA/DD. (d) MOEA/D-PBI. (e) MOEA/D-Tch. (f) MOEA/D-WS. (g) MOEA/D-IPBI (
$\theta = 0.1$
). (h) NSGA-II. (i) MOEA/D-IPBI (
$\theta = 1$
). (j) MOEA/D-IPBI (
$\theta = 5$
).
Fig. 15. - Experimental results on the ten-objective DTLZ
$2^{-1}$
 test problem. (a) NSGA-III. (b) 
$\theta $
-DEA. (c) MOEA/DD. (d) MOEA/D-PBI. (e) MOEA/D-Tch. (f) MOEA/D-WS. (g) MOEA/D-IPBI (
$\theta = 0.1$
). (h) NSGA-II. (i) MOEA/D-IPBI (
$\theta = 1$
). (j) MOEA/D-IPBI (
$\theta = 5$
).
Fig. 15.

Experimental results on the ten-objective DTLZ2^{-1} test problem. (a) NSGA-III. (b) \theta -DEA. (c) MOEA/DD. (d) MOEA/D-PBI. (e) MOEA/D-Tch. (f) MOEA/D-WS. (g) MOEA/D-IPBI (\theta = 0.1 ). (h) NSGA-II. (i) MOEA/D-IPBI (\theta = 1 ). (j) MOEA/D-IPBI (\theta = 5 ).

Fig. 16. - Experimental results on the ten-objective WFG
$9^{-1}$
 test problem. (a) NSGA-III. (b) 
$\theta $
-DEA. (c) MOEA/DD. (d) MOEA/D-PBI. (e) MOEA/D-Tch. (f) MOEA/D-WS. (g) MOEA/D-IPBI (
$\theta = 0.1$
). (h) NSGA-II. (i) MOEA/D-IPBI (
$\theta = 1$
). (j) MOEA/D-IPBI (
$\theta = 5$
).
Fig. 16. - Experimental results on the ten-objective WFG
$9^{-1}$
 test problem. (a) NSGA-III. (b) 
$\theta $
-DEA. (c) MOEA/DD. (d) MOEA/D-PBI. (e) MOEA/D-Tch. (f) MOEA/D-WS. (g) MOEA/D-IPBI (
$\theta = 0.1$
). (h) NSGA-II. (i) MOEA/D-IPBI (
$\theta = 1$
). (j) MOEA/D-IPBI (
$\theta = 5$
).
Fig. 16.

Experimental results on the ten-objective WFG9^{-1} test problem. (a) NSGA-III. (b) \theta -DEA. (c) MOEA/DD. (d) MOEA/D-PBI. (e) MOEA/D-Tch. (f) MOEA/D-WS. (g) MOEA/D-IPBI (\theta = 0.1 ). (h) NSGA-II. (i) MOEA/D-IPBI (\theta = 1 ). (j) MOEA/D-IPBI (\theta = 5 ).

A large diversity of obtained solutions by NSGA-II in Figs. 15(h) and 16(h) also explains good evaluation results by the IGD indicator in Table VII. Fig. 16(h) also shows low convergence ability of NSGA-II. In Fig. 16, the Pareto front satisfies -2i-1\le f_{i} (x)\le -1 for i = 1, 2, \ldots , 10 . However, some solutions in Fig. 16(h) have objective values close to 0.

B. Future Research Topics: Test Problems

As reported in [27]–​[32] on evolutionary many-objective optimization, very good results are obtained by weight vector-based algorithms such as NSGA-III, \theta -DEA, and MOEA/DD on the DTLZ1-4 and WFG4-9 problems in Table III. This is because the shapes of the Pareto fronts of these test problems are consistent with the shape of the distribution of the weight vectors in these algorithms. The DTLZ1-4 and WFG4-9 problems have the following Pareto fronts in the normalized objective space:\begin{equation} \sum _{i=1}^{M}\left ({y_{i}^{_{*}} }\right )^{k} = 1~{\mathrm{ and}}~y_{i}^{_{*}} \ge 0~{\mathrm{ for}}~i = 1, 2, \ldots , M \end{equation} View SourceRight-click on figure for MathML and additional features. where {k} \,\, = \,\, 1 (DTLZ1) and {k} \,\, = \,\, 2 (DTLZ2-4 and WFG4-9).

One question for future research is the similarity of these test problems to real-world many-objective problems. In other words, the question is their generality as test problems. We need much more research to answer this question. However, we can say that the Pareto front in (24) is very special in the following sense: an arbitrarily selected (M-1 ) objectives can be simultaneously optimized. For example, the Pareto front in (24) includes a Pareto optimal solution (1, 0, 0, \ldots , 0 ) where all objectives except for the first one simultaneously have their individual optimal values. This means that an ({M} -1 )-objective problem created by removing an arbitrarily selected single objective from any of the DTLZ1-4 and WFG4-9 problems always has a single absolutely optimal solution (0, 0, \ldots , 0 ). That is, the ideal point of the ({M} -1 )-objective problem is a feasible solution, which dominates all the other feasible solutions. This feature sounds very strange since it is not likely that (M-1 ) objectives of a real-world {M} -objective problem can be simultaneously optimized with no conflict among them.

The minus versions of DTLZ1-4 and WFG4-9 do not have this strange feature. However, they may have some other strange features because their original versions have the above-mentioned feature. For example, in the DTLZ1-4^{-1} problems, optimization of a single objective always leads to the worst values of all the other objectives. The Pareto optimal solution (-3.5, 0, 0, \ldots , 0 ) of DTLZ2^{-1} is an example of such a single-objective optimization result where the optimization of the first objective leads to the worst values for all the other objectives.

Our experimental results show that the consistency between the shape of the Pareto front and the shape of the distribution of the weight vectors has a large effect on the performance of weight vector-based algorithms. This was demonstrated by Jain and Deb [44] for NSGA-III on the three-objective and five-objective inverted DTLZ1 problems. Our experimental results also suggest that the size of the Pareto front may have a large effect on the performance of EMO algorithms on many-objective problems. In Fig. 17, we show the Pareto fronts of the two-objective DTLZ2 and DTLZ2^{-1} problems together with their feasible region and randomly generated 100 initial solutions. In Fig. 17, DTLZ2^{-1} is shown as the maximization problem of DTLZ2: Max-DTLZ2. Thus the two problems (DTLZ2 and its maximization version) have exactly the same feasible region. As we can see from Fig. 17, DTLZ2 has a small Pareto front in comparison with the spread of the randomly generated initial solutions. It is likely that a good solution set can be obtained for DTLZ2 by pulling those initial solutions toward the ideal point. That is, the convergence of solutions toward the Pareto front is important for DTLZ2.

Fig. 17. - Pareto fronts of the two-objective DTLZ2 and its maximization version Max-DTLZ2 which is equivalent to the two-objective DTLZ
$2^{-1}$
.
Fig. 17. - Pareto fronts of the two-objective DTLZ2 and its maximization version Max-DTLZ2 which is equivalent to the two-objective DTLZ
$2^{-1}$
.
Fig. 17.

Pareto fronts of the two-objective DTLZ2 and its maximization version Max-DTLZ2 which is equivalent to the two-objective DTLZ2^{-1} .

However, in DTLZ2^{-1} in Fig. 17, not only the convergence of solutions but also the diversification of solutions is very important for finding a set of well-distributed solutions over the entire Pareto front. Moreover, the improvement in the diversity of solutions in Fig. 17 has a positive effect on the convergence of solutions toward the Pareto front of DTLZ2^{-1} (whereas it has a negative effect for DTLZ2). For example, the crowding distance-based fitness evaluation tends to increase the range of objective values for each objective in the population. In Fig. 17, the maximum range for each objective is obtained by the two extreme solutions on the edge of the Pareto front of DTLZ2^{-1} [i.e., one extreme solution with the maximum value of {f} _{1} (\boldsymbol {x} ) and the other with the maximum value of {f} _{2} (\boldsymbol {x} ) of Max-DTLZ2]. These two extreme solutions will help the convergence of other solutions toward the Pareto front of DTLZ2^{-1} . However, they have a negative effect on the convergence of solutions toward the Pareto front of the original DTLZ2. The difference in the effects of the diversity improvement efforts on the convergence of solutions toward the Pareto front between DTLZ2 and DTLZ2^{-1} in Fig. 17 explains the difference of the performance of NSGA-II between DTLZ1-4 and DTLZ1-4^{-1} in Tables III–​VII. NSGA-II is evaluated as the worst among the examined eight algorithms for DTLZ1-4 by the hypervolume in Table III and also by the IGD indicator in Table VI. However, it is evaluated as the best algorithm for DTLZ1-4^{-1} by IGD in Table VII.

Let us briefly explain why good solutions can be obtained for many-objective DTLZ, WFG, DTLZ^{-1} , and WFG^{-1} test problems as shown in Figs. 5, 15, and 16. In all of the DTLZ1-4 and DTLZ1-4^{-1} problems and some of the WFG1-9 and WFG1-9^{-1} problems, decision variables are separable into distance variables and position variables. In those problems, any changes only in the distance variables simply modify only the distance of a solution from the Pareto front. When only position variables are changed, only the position is changed without changing the value of the distance function. For illustrating these properties of most test problems based on their separable decision variables, first we randomly generate five initial solutions. The generated five initial solutions are shown by open circles in Fig. 18. Then we generate a new solution by randomly changing the values of all distance variables of each initial solution (i.e., by replacing the current values of the distance variables with randomly generated values within their domains). This is iterated 100 times to generate 100 solutions for each initial solution. In this manner, we generate 500 solutions. Fig. 18 shows the generated solutions for the two-objective DTLZ1-4 and WFG1-9 problems [and DTLZ1 with five and ten objectives in Fig. 18(a2) and (a3)]. We also perform the same experiment by changing the values of all position variables (see Fig. 19).

Fig. 18. - Generated solutions by changing distance variables. (a1) 2-Obj. DTLZ1. (a2) 5-Obj. DTLZ1. (a3) 10-Obj. DTLZ1. (b) 2-Obj. DTLZ2. (c) 2-Obj. DTLZ3. (d) 2-Obj. DTLZ4. (e) 2-Obj. WFG1. (f) 2-Obj. WFG2. (g) 2-Obj. WFG3. (h) 2-Obj. WFG4. (i) 2-Obj. WFG5. (j) 2-Obj. WFG6. (k) 2-Obj. WFG7. (l) 2-Obj. WFG8. (m) 2-Obj. WFG9.
Fig. 18. - Generated solutions by changing distance variables. (a1) 2-Obj. DTLZ1. (a2) 5-Obj. DTLZ1. (a3) 10-Obj. DTLZ1. (b) 2-Obj. DTLZ2. (c) 2-Obj. DTLZ3. (d) 2-Obj. DTLZ4. (e) 2-Obj. WFG1. (f) 2-Obj. WFG2. (g) 2-Obj. WFG3. (h) 2-Obj. WFG4. (i) 2-Obj. WFG5. (j) 2-Obj. WFG6. (k) 2-Obj. WFG7. (l) 2-Obj. WFG8. (m) 2-Obj. WFG9.
Fig. 18.

Generated solutions by changing distance variables. (a1) 2-Obj. DTLZ1. (a2) 5-Obj. DTLZ1. (a3) 10-Obj. DTLZ1. (b) 2-Obj. DTLZ2. (c) 2-Obj. DTLZ3. (d) 2-Obj. DTLZ4. (e) 2-Obj. WFG1. (f) 2-Obj. WFG2. (g) 2-Obj. WFG3. (h) 2-Obj. WFG4. (i) 2-Obj. WFG5. (j) 2-Obj. WFG6. (k) 2-Obj. WFG7. (l) 2-Obj. WFG8. (m) 2-Obj. WFG9.

Fig. 19. - Generated solutions by changing position variables. (a1) 2-Obj. DTLZ1. (a2) 5-Obj. DTLZ1. (a3) 10-Obj. DTLZ1. (b) 2-Obj. DTLZ2. (c) 2-Obj. DTLZ3. (d) 2-Obj. DTLZ4. (e) 2-Obj. WFG1. (f) 2-Obj. WFG2. (g) 2-Obj. WFG3. (h) 2-Obj. WFG4. (i) 2-Obj. WFG5. (j) 2-Obj. WFG6. (k) 2-Obj. WFG7. (l) 2-Obj. WFG8. (m) 2-Obj. WFG9.
Fig. 19. - Generated solutions by changing position variables. (a1) 2-Obj. DTLZ1. (a2) 5-Obj. DTLZ1. (a3) 10-Obj. DTLZ1. (b) 2-Obj. DTLZ2. (c) 2-Obj. DTLZ3. (d) 2-Obj. DTLZ4. (e) 2-Obj. WFG1. (f) 2-Obj. WFG2. (g) 2-Obj. WFG3. (h) 2-Obj. WFG4. (i) 2-Obj. WFG5. (j) 2-Obj. WFG6. (k) 2-Obj. WFG7. (l) 2-Obj. WFG8. (m) 2-Obj. WFG9.
Fig. 19.

Generated solutions by changing position variables. (a1) 2-Obj. DTLZ1. (a2) 5-Obj. DTLZ1. (a3) 10-Obj. DTLZ1. (b) 2-Obj. DTLZ2. (c) 2-Obj. DTLZ3. (d) 2-Obj. DTLZ4. (e) 2-Obj. WFG1. (f) 2-Obj. WFG2. (g) 2-Obj. WFG3. (h) 2-Obj. WFG4. (i) 2-Obj. WFG5. (j) 2-Obj. WFG6. (k) 2-Obj. WFG7. (l) 2-Obj. WFG8. (m) 2-Obj. WFG9.

In Fig. 18, the generated solutions from the same parent are on the same line except for WFG7 and WFG9. Since this also holds for many-objective problems from the DTLZ and WFG problem formulations, the generated solutions by changing distance variables can be compared by the Pareto dominance relation. This explains why good results are obtained by NSGA-II for some many-objective test problems. However, this is totally different from a general case of many-objective optimization where almost all solutions are nondominated with each other. Moreover, each test problem in DTLZ and WFG has only a single distance function independent of the number of objectives. By optimizing the single distance function, we can obtain Pareto optimal solutions. This means that the search for Pareto optimal solutions is single-objective optimization independent of the number of objectives.

Fig. 19 shows the generated solutions by changing position variables. Except for WFG2, generated solutions from the same solution have the same distance from the Pareto front. This also holds for many-objective problems from the DTLZ and WFG problem formulations. If a solution is Pareto optimal (i.e., if it is on the Pareto front), all solutions generated by changing position variables are also Pareto optimal. Thus, we can easily increase the diversity of solutions by changing position variables without deteriorating their convergence. From these discussions on Figs. 18 and 19, we can see that most test problems in DTLZ and WFG are not difficult as many-objective problems. This observation is consistent with the reported results by the weight vector-based algorithms in [27]–​[32] where very good solutions were obtained (which are almost the same as a reference point set for the IGD calculation as shown in Fig. 5). This is also consistent with our experimental results by NSGA-II (e.g., the best results are obtained by NSGA-II for some of the WFG problems in Table VI with the IGD indicator).

Our observation from Figs. 18 and 19 also explains the features of some weight vector-based algorithms. Since the Pareto dominance relation holds among solutions generated by changing distance variables, Pareto dominance-based fitness evaluation is used in some weight vector-based algorithms (whereas it does not work well on many-objective problems in general). For the same reason, the convergence improvement is not difficult. Thus a large penalty value (i.e., \theta = 5 ) is used in the PBI function even for many-objective problems (whereas the PBI function with a large penalty value does not work well on many-objective problems in general). These discussions suggest that the recent development of weight vector-based algorithms is overspecialized for DTLZ and WFG with respect to not only the shape of the distribution of the weight vectors but also their fitness evaluation mechanisms (in other words, their convergence-diversification balance).

As we have explained in this paper, the DTLZ and WFG problems are test problems with special characteristic features. Thus an important future research topic is the creation of a wide variety of test problems with respect to the shape of the Pareto front, the size of the Pareto front, the relation among decision variables, and the relation among objectives. The size of the Pareto front can be rephrased as the shape of the feasible region in the objective space.

C. Future Research Topics: Algorithm Developments

Our experimental results showed that good results were obtained when the shape of the distribution of weight vectors is the same as or similar to the shape of the Pareto front. In general, we do not know the shape of the Pareto front. It is much more difficult to find the exact shape of the Pareto front of a many-objective problem than the case of two or three objectives. So it may be desirable for many-objective optimizers to have robust search ability with respect to the shape of the Pareto front. A simple idea is to simultaneously use multiple sets of weight vectors with different distributions. For example, it may be a good idea to simultaneously use both the PBI and IPBI functions in a single MOEA/D algorithm. In [50], the simultaneous use of the weighted sum and the weighted Tchebycheff was examined. Then it was shown that better results were obtained from their simultaneous use than their individual use (i.e., MOEA/D-WS and MOEA/D-Tch) on many-objective knapsack problems.

Another idea is the adaptation of the distribution of weight vectors to the shape of the Pareto front. This idea has been suggested for MOEA/D in pa\lambda -MOEA/D [51]. Adaptive NSGA-III (A-NSGA-III) [44] can be viewed as a representative algorithm in such a research direction. A-NSGA-III has simple mechanisms for weight vector deletion and creation. Its performance has not been evaluated on many-objective problems with eight or more objectives (i.e., many-objective problems for which the two-layered approach was used to generate weight vectors). Development of adaptive many-objective algorithms seems to be an interesting future research direction. However, before implementing an efficient and effective adaptation mechanism of the weight vector distribution, we may need to further examine the behavior of weight vector-based algorithms on a wide variety of many-objective test problems under various settings such as a large number of weight vectors together with a large population, a large number of weight vectors together with a small population, and a small number of weight vectors together with a large population. Analyzing multiobjective search behavior under these settings may help us to understand the necessity of weight vector deletion and creation.

SECTION VI.

Conclusion

In this paper, we clearly showed the similarity between the shape of the distribution of the weight vectors in weight vector-based many-objective algorithms and the shape of the Pareto fronts of frequently used many-objective test problems (i.e., DTLZ1-4 and WFG4-9). For demonstrating the high sensitivity of their performance to the shape of the Pareto front, we formulated DTLZ^{-1} and WFG^{-1} problems by multiplying (−1) to each objective of the DTLZ and WFG problems. Whereas DTLZ1-4 and WFG4-9 have triangular shape Pareto fronts, DTLZ1-4^{-1} and WFG4-9^{-1} have rotated triangular shape Pareto fronts. We demonstrated through computational experiments that totally different performance comparison results among different algorithms were obtained from the original test problems and their minus versions. Whereas the best results were obtained by \theta -DEA and MOEA/DD on the original DTLZ and WFG problems, they were outperformed by MOEA/D-WS and MOEA/D-IPBI on their minus versions with 3, 5, 8, and 10 objectives. Those high performance weight vector-based algorithms on the original DTLZ and WFG problems were also outperformed by NSGA-II on some of the minus versions even when they have ten objectives.

One difficulty of recently developed weight vector-based algorithms is the lack of appropriate handling of reference lines outside the Pareto front (i.e., with no intersection with the Pareto front). This difficulty can be rephrased as the lack of appropriate criteria for choosing the second solution for each reference line. This difficulty was demonstrated by our experimental results on DTLZ^{-1} and WFG^{-1} where NSGA-III with random selection of the second solution outperformed \theta -DEA and MOEA/DD with the selection of the second best solution. This difficulty is not clear when we apply them to frequently used many-objective test problems such as DTLZ1-4 and WFG4-9. This is because all reference lines are inside the Pareto fronts of those test problems thanks to their triangular shape Pareto fronts. In general, the above-mentioned difficulty is severe when many reference lines are outside the Pareto front. In this sense, many-objective test problems with degenerate Pareto fronts are difficult for weight vector-based algorithms. This discussion is consistent with Table III where the best results were obtained by NSGA-II for all test problems of WFG3 and also with Table VI where good results were not obtained by \theta -DEA for any test problem of WFG3. Whereas WFG3 is not a degenerate test problem, the nondegenerate part of its Pareto front is very small (see [38] for the shape of the Pareto front of WFG3). Thus many weight vectors have no intersection with the Pareto front. The use of a wide variety of test problems with various Pareto front shapes will prevent the development of weight vector-based algorithms from being overspecialized for a special class of many-objective test problems such as DTLZ1-4 and WFG4-9. The use of a wide variety of test problems will encourage the development of appropriate handling mechanisms of reference lines outside the Pareto front including their online adaptation.

We also suggested that the DTLZ and WFG test problems are not difficult even when they have many objectives due to their special features in the test problem formulations. As a result, weight vector-based algorithms developed for those problems have fitness evaluation mechanisms which are not always suitable for many-objective optimization such as Pareto dominance and the emphasis on the minimization of the distance from the nearest reference line (e.g., the use of a large penalty value for the distance {d} _{2} in the PBI function). These discussions also clearly show the necessity of a wide variety of many-objective test problems.

Our computational experiments explained the following reasons for high performance of recently proposed weight vector-based algorithms on many-objective DTLZ1-4 and WFG4-9 test problems.

  1. Triangular shape Pareto front of each test problem, which is the same as or similar to the shape of the distribution of the weight vectors.

  2. Relatively small size Pareto front of each test problem in comparison with the feasible region in the objective space, which is suitable for multiobjective search by pulling solutions toward the ideal point using the weight vectors.

  3. Easy convergence and easy diversification of solutions due to separable decision variables, which make it possible to focus on the uniformity of obtained solutions over the Pareto front.

Thanks to these special features of frequently used many-objective test problems, uniformly distributed and well converged solutions can be obtained over the entire Pareto front even for the case of many objectives. However, these features are not likely to hold in real-world applications. Our computational experiments showed that high performance of recently proposed weight vector-based algorithms on many-objective test problems was deteriorated by changing the shape and the size of the Pareto front of each test problem. In the design of an MOEA/D-based algorithm for a many-objective problem at hand, weight vectors should be carefully specified based on the shape and the size of the Pareto front. A scalarizing function (i.e., a single-objective optimization problem) should be carefully defined for each weight vector depending on the difficulty of the convergence and the diversification in order to strike a good balance between them. However, these features (i.e., Pareto front shape, Pareto front size, convergence difficulty, and diversification difficulty) are usually unknown in real-world applications. Thus we may need an adaptation mechanism for the weight vectors and the scalarizing function. The framework of MOEA/D continues to be useful for developing such an adaptive many-objective algorithm. This is because MOEA/D has a large flexibility in weight vector specification and scalarizing function definition. We can use any set of weight vectors as well as any scalarizing function in MOEA/D-based many-objective algorithms. In an adaptive many-objective algorithm, the distribution of weight vectors will be adjusted to the shape and the size of the Pareto front, and the search mechanism for each weight vector will be adjusted by changing its scalarizing function depending on the difficulty of the convergence and the diversification.

References

References is not available for this document.