Introduction
Various multiobjective evolutionary algorithms (MOEAs) have been proposed to effectively solve multiobjective optimization problems (MOPs). However, most of these algorithms have been evaluated and applied to MOPs with only two or three objectives [1], [2], despite the fact that optimization problems involving a high number of objectives indeed appear widely in real-world applications [3]–[5]. In the literature, such MOPs that have more than three objectives are often termed many-objective optimization problems (MaOPs) [6], [7].
Unfortunately, experimental and analytical results [8], [9] have indicated that MaOPs would pose serious difficulties to the existing MOEAs, particularly for the popular Pareto-dominance-based MOEAs, e.g., nondominated sorting genetic algorithm II (NSGA-II) [10] and strength Pareto evolutionary algorithm 2 (SPEA2) [11]. The primary reason for this is that the proportion of nondominated solutions in a population increases exponentially with the number of objectives and it would lead to the severe loss of Pareto-dominance-based selection pressure toward the Pareto front (PF). In the existing research, there generally exist three types of techniques to overcome the shortage of Pareto-dominance-based MOEAs.
The idea of the first type of approaches is to adopt new preference relations (relaxed forms of Pareto dominance or other ranking schemes) that could induce finer grain order in the objective space. So far, many alternative preference relations have been proposed for many-objective optimization, such as average and maximum ranking [12], favor relation [13], fuzzy Pareto dominance [14], expansion relation [15], preference order ranking [16],
The second type of approach intends to improve the diversity maintenance mechanism. The consideration is that since Pareto dominance could not exert sufficient selection pressure toward the PF in a high-dimensional objective space, then the selection should be almost solely based on the diversity, which is generally regarded as the secondary selection operator in MOEAs. However, most existing diversity maintenance criteria, e.g., crowding distance [10], prefer dominance resistant [8] solutions, which would bias the search toward the solutions with the poor proximity to the global PF, although these solutions may present good diversity over the objective space. Thus, the diversity preservation in many-objective scenario requires some care to avoid or weaken this phenomenon. Compared to the first type of research, the work along this direction [24]–[26] has received much less attention.
In many-objective optimization, it is indeed very difficult for a MOEA to emphasize both convergence and diversity within a single population or archive. Motivated by this, the third type of approaches maintains two archives for convergence and diversity separately during the evolutionary search, and at least one of the two archives still uses Pareto dominance. The related work can be referred in [27] and [28].
Unlike Pareto-dominance-based MOEAs, another two classes of MOEAs, indicator- and decomposition-based MOEAs, have not been criticized much on the selection pressure produced in a high-dimensional objective space. However, they would encounter their own difficulties when handling MaOPs.
The indicator-based MOEAs aim to optimize a performance indicator that provides a desired ordering among sets that represent PF approximations. The hypervolume (HV) [1] may be the most popular performance indicator ever adopted due to its nice theoretical properties [29]. There have been a few well-established indicator-based MOEAs [30]–[32] available in the literature that employ the HV as the selection criterion. Nevertheless, the computational cost of HV grows exponentially with the number of objectives increasing. To address this issue, one of the alternative strategies is to estimate the rankings of solutions induced by the HV indicator without computing the exact indicator values [33]. Another strategy is to replace the HV with the other indicators that are much less computationally expensive but also have good theoretical properties, e.g.,
The decomposition-based MOEAs generally aggregate the objectives of a MOP into an aggregation function using a unique weight vector, and a set of weight vectors (or directions) will generate multiple aggregation functions, each of which defines a single-objective optimization problem. The diversity of the population is implicitly maintained by specifying a good spread of the weight vectors in the objective space. MOEA based on decomposition (MOEA/D) [39], [40] and multiple single-objective Pareto sampling (MSOPS) [41] are two most typical decomposition-based MOEAs. Very recently, Yuan et al. [42] proposed a new framework based on NSGA-II for many-objective optimization, called ensemble fitness ranking (EFR), which is more general than MSOPS. In terms of the form, EFR can be viewed as an extension to average ranking and maximum ranking [12]. The significant difference lies in that EFR employs more general fitness functions (aggregation functions) instead of objective functions. However, there is still deficiency for decomposition-based MOEAs to conduct many-objective search. Take MOEA/D for instance, it normally could approach the PF very well even if in the high-dimensional objective space, but it struggles to maintain the diversity in many-objective optimization, usually failing to achieve a good coverage of PF [20], [26], [43]. The reason is largely due to the nature of contour lines for the adopted aggregation functions, which will be further explained in Section III. The other decomposition-based MOEAs, e.g, MSOPS and EFR, suffer from a similar problem, since they also drive the search using aggregation functions. Note that, as a major framework to design MOEAs, MOEA/D has spawned a large amount of research work (see [44]–[47]), but almost all the newly proposed MOEA/D variants are yet to be validated on MaOPs to evaluate their suitability in many-objective optimization. Moreover, several recent decomposition-based MOEAs, e.g., MOEA/D-M2M [47] and NSGA-III [26], adopt new decomposition scheme, which decomposes the objective space into subregions without using any aggregation function. Although these algorithms address the difficulty in diversity maintenance to some extent, they struggle to converge to the PF in many-objective optimization as the Pareto dominance is still relied on to provide the selection pressure.
This paper mainly focuses on the decomposition-based MOEAs that use aggregation functions to promote the convergence in many-objective optimization. Our contributions to this topic are summarized as follows.
We have proposed a better tradeoff between convergence and diversity in decomposition-based many-objective optimizers by exploiting the perpendicular distance from a solution to the weight vector in the objective space.
We have implemented this idea in MOEA/D and EFR, respectively, to improve their performance on MaOPs, resulting in two enhanced algorithms, i.e., a MOEA/D variant (MOEA/D-DU) with a distance-based updating strategy and a new EFR version (EFR-RR) with a ranking restriction scheme.
We have provided an optional online normalization procedure that can be incorporated into MOEA/D-DU and EFR-RR to effectively tackle scaling problems, which is a new practice [26], [48] and has not been extensively studied in the literature.
We have combined for the first time different existing ideas together synergistically in a single MOEA. For example, MOEA/D-DU inherits the merit of systematic sampling (the same as [26]), notion of neighborhood (the same as [40]), adaptive normalization (similar to [26] and [48]), single first replacement (the same as [48]), and the preferred order of replacement (introduced in this paper).
We have discussed about the similarities and differences between our proposals and the other existing decomposition-based approaches. Based on the extensive experimental studies, we have also given possible explanations on why some of these approaches are inferior to ours in many-objective optimization.
Based on our experimental studies, we have suggested a practice to compare the performance of MOEAs (with or without sophisticated normalization) on benchmark problems more fairly and reasonably.
The rest of this paper is organized as follows. In Section II, the background knowledge is introduced. In Section III, the basic idea is given based on the analysis of the drawback of original MOEA/D and EFR. Section IV describes in detail how to implement the basic idea to enhance MOEA/D and EFR in many-objective optimization, respectively. Section V presents the test problems, performance metrics, and algorithm settings used for performance comparison. Section VI experimentally investigates the performance of enhanced algorithms from two aspects. In Section VII, the extensive experiments are carried out to compare the enhanced algorithms with state-of-the-art methods. Finally, Section VIII concludes this paper.
Preliminaries and Background
In this section, some basic concepts in multiobjective optimization are first given. Then, we briefly introduce several most recent proposals in the literature that are somewhat similar to ours.
A. Basic Concepts
A MOP can be mathematically defined as \begin{align} \min \quad&\mathbf {f}(\,\mathbf {\mathbf {x}}) = \left ({\,f_{1}(\mathbf {x}), f_{2}(\mathbf {x}), \ldots , f_{m}(\mathbf {x})}\right )^{\mathrm {T}} \notag \\ \text {subject to} \quad&\mathbf {x} \in \Omega \subseteq \mathbb {R}^{n} \quad \end{align}
Definition 1:
A vector
Definition 2:
For a given MOP, a decision vector
Definition 3:
For a given MOP, the Pareto set, PS, is defined as \begin{equation} {\rm PS} = \left \{{\,\mathbf {x} \in \Omega | \mathbf {x}~\text {is Pareto optimal}}\right \}. \end{equation}
Definition 4:
For a given MOP, the PF is defined as \begin{equation} {\rm PF} = \left \{{\,\mathbf {f}(\mathbf {x}) \in \Theta | \mathbf {x} \in {\rm PS}}\right \}. \end{equation}
Definition 5:
For a given MOP, the ideal point
Definition 6:
For a given MOP, the nadir point
The goal of multiobjective optimization is to find a set of nondominated objective vectors that are close to the PF (convergence), and also enable them to distribute well over the PF (diversity).
B. Related Approaches
In this section, we would review several representative decomposition-based approaches relating to this paper.
Qi et al. [45] proposed an improved MOEA/D with the adaptive weight vector adjustment (MOEA/D-AWA), which mainly enhances MOEA/D in two aspects. First, MOEA/D-AWA uses a new weight vector initialization method based on the geometric analysis of the original Tchebycheff function. Second, MOEA/D-AWA adopts an adaptive weight vector adjustment strategy to deal with problems having complex PFs. This strategy adjusts weights periodically so that weights of subproblems can be redistributed adaptively to obtain better uniformity of solutions.
Li et al. [46] proposed to use a stable matching (STM) model to coordinate the selection process in MOEA/D, and the resulting MOEA/D variant is referred to as MOEA/D-STM. In MOEA/D-STM, the subproblems and solutions are regarded as two sets of agents. A subproblem prefers solutions which can lower its aggregation function values, while a solution prefers subproblems whose corresponding direction vectors are close to it. To balance the preference of subproblems and solutions, a matching algorithm is used to assign a solution to each subproblem in order to balance the convergence and diversity of the evolutionary search. Unlike MOEA/D, MOEA/D-STM uses the generational scheme.
Deb and Himanshu [26] suggested a reference-point-based many-objective NSGA-II, called NSGA-III. Similar to NSGA-II, NSGA-III still employs the Pareto nondominated sorting to partition the population into a number of nondomination fronts. In the last accepted front, instead of using the crowding distance, the solutions are selected based on a niche-preservation operator, where the solutions associated with less crowded reference line have higher priority to be chosen. Overall, NSGA-III emphasizes the nondominated solutions yet close to the reference lines of a set of supplied reference points. A sophisticated online normalization procedure is incorporated into NSGA-III to effectively handle scaled test problems.
Just around the time when this paper was originally submitted, the following two proposals have been available in the literature.
Asafuddoula et al. [48] proposed an improved decomposition-based evolutionary algorithm (I-DBEA) for many-objective optimization. In I-DBEA, the child solution attempts to enter the population via replacement only if it is nondominated with respect to the individuals in the population. The child solution competes with all the solutions in the population in a random order until it makes a successful replacement or all the solutions have been competed. I-DBEA also includes an online normalization procedure similar to that in NSGA-III.
Wang et al. [49] presented a MOEA/D variant with the global replacement (MOEA/D-GR). Once a new solution is produced in MOEA/D-GR, it is associated with a subproblem on which it achieves the minimum aggregation function value. Then, a number of closest subproblems to the associated subproblem are selected to form the replacement neighborhood, and the new solution attempts to replace the current solutions to these selected subproblems.
Note that, MOEA/D-STM and MOEA/D-GR were studied and verified only on 2- and 3-objective problems. MOEA/D-AWA was studied on MaOPs, but the problems were just limited to degenerate problems. NSGA-III and I-DBEA were designed specially for many-objective optimization. Moreover, although NSGA-III and I-DBEA use the decomposition-based idea, they still use the Pareto dominance instead of aggregation functions to control the convergence, which differ significantly from the other mentioned decomposition-based approaches.
Basic Idea
The Tchebycheff function is one of the most common type of aggregation functions used in decomposition-based MOEAs. In this paper, we adopt a modified version of the Tchebycheff function. Let \begin{equation} \mathcal {F}_{j}(\mathbf {x}) = \max _{k=1}^{m}\left \{{\frac {1}{\lambda _{j, k}}\left |{\,f_{k}(\mathbf {x}) - z_{k}^{*}}\right |}\right \} \end{equation}
This form of Tchebycheff function has two advantages over the original one [39]. First, the uniformly distributed weight vectors lead to the uniformly distributed search directions in the objective space. Second, each weight vector corresponds to a unique solution located on PF. The proof can be found in [45]. The two advantages alleviate the difficulty in diversity preservation to some extent.
However, even (4) is not perfect in practice. Ideally, if the optimal solution is obtained for each
Both situations mentioned above could be attributed to the fact that the solution which is distant to
Given all these, we are motivated to consider not only the aggregation function value of a solution but also its distance to the corresponding weight vector in MOEA/D and EFR. This practice is expected to force the solutions to stay close to the weight vectors, and explicitly maintain the desired distribution of solutions in the evolutionary process, leading to the better balance between convergence and diversity in many-objective optimization. It is worth mentioning that the penalty-based boundary intersection function [39] implicitly considers the closeness of a solution to the weight vector to certain extent, but it still has the problem mentioned above. In this paper, without loss of generality, we only consider the modified Tchebycheff function defined in (4).
Suppose that \begin{equation} d_{j,2}(\mathbf {x}) = \left \|{\mathbf {f}(\mathbf {x}) - \mathbf {z}^{*} - d_{j,1}(\mathbf {x})\left ({ \boldsymbol {\lambda } _{j} \big / \left \|{ \boldsymbol {\lambda } _{j}}\right \|}\right )}\right \| \end{equation}
\begin{equation} d_{j,1}(\mathbf {x}) = \left \|{\left ({\,\mathbf {f}(\mathbf {x}) - \mathbf {z}^{*}}\right )^{\mathrm {T}} \boldsymbol {\lambda } _{j}}\right \| \Big / \left \|{ \boldsymbol {\lambda } _{j}}\right \|. \end{equation}
Illustration of the perpendicular distance from the solution to the weight vector in the 2-D objective space.
Proposed Algorithms
In this section, two enhanced algorithms, i.e., MOEA/D-DU and EFR-RR, are described in detail in Sections IV-A and IV-B, respectively. Section IV-C provides an optional normalization procedure that can be incorporated into the proposed algorithms. Section IV-D briefly analyzes the computational complexity of one generation of MOEA/D-DU and EFR-RR. In Section IV-E, we discuss the similarities and differences between the proposed algorithms and several existing ones in the literature.
A. Enhancing MOEA/D
The framework of the proposed MOEA/D-DU is depicted in Algorithm 1. First, a set of uniformly spread weight vectors
Algorithm 1 Framework of MOEA/D-DU
Generate a set of weight vectors
Initialize the population
Initialize the ideal point
for
end for
while the termination criterion is not satisfied do
for
if
else
end if
Randomly select an index
UpdateIdealPoint(
UpdateCurrentPopulation(
end for
end while
return all the nondominated solutions in
The updating strategy (step 17 of Algorithm 1) is the characteristic procedure in MOEA/D-DU, which is significantly different from the original MOEA/D. This strategy is presented in detail in Algorithm 2, which works as follows. Once a new solution
Algorithm 2 UpdateCurrentPopulation($\mathbf {y}, \mathbf {z}^{\ast }, \Lambda , P, K$
)
for
Compute the perpendicular distance from
end for
Select
for
if
return
end if
end for
B. Enhancing EFR
EFR [42] adopts the generational scheme, and its framework is based on NSGA-II but with significant differences in the environmental selection phase.
At the \begin{equation} R_{g}(\mathbf {x}) = \min _{j = 1}^{N} r_{j}(\mathbf {x}). \end{equation}
From the above description, it can be seen that all the solutions in \begin{equation} R_{g}(\mathbf {x}) = \min _{j \in B(\mathbf {x})} r_{j}(\mathbf {x}). \end{equation}
In Algorithm 3, we describe the procedure of maximum ranking in EFR-RR. Further, we summarize the whole procedure of EFR-RR in Algorithm 4.
Algorithm 3 MaximumRanking $(U_{t}, K)$
for each solution
Compute the set
for each
end for
end for
for
Sort
for
if
end if
end for
end for
return
Algorithm 4 Framework of EFR-RR
Generate a set of weight vectors
Initialize the population
Initialize the ideal point
while the termination criterion is not met do
while
end while
Randomly shuffle
end while
C. Optional Normalization Procedure
Normalization is particularly useful for an algorithm to solve problems having a PF whose objective values are differently scaled. For example, Zhang and Li [39] showed that even a naive normalization procedure can significantly improve the performance of MOEA/D on scaled test problems.
In essence, the task of normalization is to estimate \begin{equation} \tilde {f}_{i}(\mathbf {x}) = \left ({\,f_{i}(\mathbf {x}) - z_{i}^{\ast }}\right ) \Big / \left ({z_{i}^{\text {nad}} - z_{i}^{\ast }}\right ). \end{equation}
In this paper, we provide an online normalization procedure similar to that in NSGA-III, which can be incorporated into the proposed algorithms to deal with scaled test problems.
The difference between normalization presented here and that in NSGA-III lies in that we use a slightly different achievement scalarizing function to identify extreme points. Suppose \begin{equation} {\rm ASF}\left ({\mathbf {x}, \mathbf {w}_{j}}\right ) = \max _{i=1}^{m}\left \{{\frac {1}{w_{j,i}}\left |{\frac {f_{i}(\mathbf {x}) - z_{i}^{\ast }}{z_{i}^{\text {nad}} - z_{i}^{\ast }}}\right |}\right \} \end{equation}
It is worth pointing out that
D. Computational Complexity
For MOEA/D-DU, the major computational costs are in the updating procedure depicted in Algorithm 2. Steps 1–3 require
As for EFR-RR, the major computation costs lie in the calculation of global ranks. First, the computation of perpendicular distances totally requires
E. Discussion
After describing the details of MOEA/D-DU and EFR-RR, we would like to discuss main similarities and differences between the proposed algorithms and the ones listed in Section II-B. It should be pointed out that all these algorithms employ a set of weight vectors to guide the search process.
1) MOEA/D-DU Versus MOEA/D-AWA:
Although both of them make sure that the uniformly spread weight vectors lead to uniformly spread search directions in the objective space, they use slightly different means to implement this. MOEA/D-AWA adopts a new weight vector initialization method, whereas MOEA/D-DU uses a modified form of aggregation function. More importantly, MOEA/D-AWA aims to adjust the weights periodically so as to obtain a better uniformity of solutions on problems having complex PFs, whereas MOEA/D-DU intends to maintain a better balance between convergence and diversity using a set of fixed weight vectors. In other words, MOEA/D-AWA emphasizes weight adjustment strategy, whereas MOEA/D-DU emphasizes a new strategy to update the population.
2) MOEA/D-DU Versus MOEA/D-STM:
Both of them use the modified version of Tchebycheff function and employ
3) MOEA/D-DU Versus I-DBEA:
Both of them enhance MOEA/D with major modifications in the updating procedure and need to compute
4) MOEA/D-DU Versus MOEA/D-GR:
Both of them conduct a careful selection of solutions that can be replaced by the new solution. However, the two algorithms have rather different original intentions. MOEA/D-GR aims to find a most suitable subproblem for the newly produced solution, and the neighborhood concept of subproblems defined in MOEA/D is still used in the replacement scheme of MOEA/D-GR. Unlike MOEA/D-GR, MOEA/D-DU does not use the neighborhood of subproblems any longer in the updating procedure. Instead, it takes a pure geometric view and just exploits
5) EFR-RR Versus NSGA-III:
Both of them use the generational scheme and split the combined population into a number of “fronts” to select solutions. However, there exists significant difference between them in how to rank solutions in the environmental selection phase. In NSGA-III, the Pareto dominance is still used to promote the convergence, and a niche-preservation operator aided by a set of weight vectors is adopted to select solutions in the last accepted front, intending to maintain the population diversity. In contrast to NSGA-III, EFR-RR promotes the convergence using the aggregation functions instead of Pareto dominance. Note that EFR-RR also performs the niching procedure with the aid of weight vectors, thus each weight vector in EFR-RR serves the purpose of promoting both convergence and diversity. Whereas in NSGA-III, each weight vector is mainly used to emphasize the diversity. Even when it comes to niching methodology, there is a distinction between NSGA-III and EFR-RR. In NSGA-III, each solution can only be associated with a sole weight vector; but in EFR-RR, each solution can be associated with several weight vectors. In Section VII-B, we would show that EFR-RR is compared favorably with NSGA-III on scaled problems.
Experimental Design
This section is devoted to the experimental design for investigating the performance of MOEA/D-DU and EFR-RR. First, the test problems and performance metrics used in the experiments are given. Then, we list eight MOEAs that are used to validate the proposed algorithms. Finally, the experimental settings adopted in this paper are provided.
A. Test Problems
For comparison purposes, we use the test problems from DTLZ [53] and WFG [54] test suites. We divide the problems into two groups.
The first group of problems are all normalized test problems including DTLZ1–DTLZ4, DTLZ7, and WFG1–WFG9. Note that the objective values of the original DTLZ7 and WFG1–WFG9 problems are scaled differently. Since their true PFs are known, we modify their objective functions as \begin{equation} f_{i} \leftarrow \left ({\,f_{i} - z_{i}^{\ast }}\right ) \Big / \left ({z_{i}^{\text {nad}} - z_{i}^{\ast }}\right ), \hspace {5pt} i = 1, 2, \ldots , m. \end{equation}
The second group of problems are all scaled test problems, including scaled DTLZ1 and DTLZ2 problems [26] and the original WFG4–WFG9 problems. The scaled DTLZ1 and DTLZ2 are the modifications of DTLZ1 and DTLZ2, respectively. To illustrate, if the scaling factor is \begin{equation} f_{i} \leftarrow 10^{i-1}f_{i}, \hspace {5pt} i = 1, 2, \ldots , m. \end{equation}
All these problems can be scaled to any number of objectives and decision variables. In this paper, we consider the number of objectives
B. Performance Metrics
The performance metrics are needed to evaluate the performance of the concerned algorithms. The inverted generational distance (IGD) [55] is one of the most widely used metrics in multiobjective scenario, providing a combined information about convergence and diversity of a solution set. However, in high-dimensional objective space, a huge number of uniformly distributed points on the PF are required to calculate IGD in a reliable manner [56]. In this paper, we use another very popular metric, HV [1], as the primary comparison criterion. The HV metric is strictly Pareto compliant [55], and its nice theoretical qualities make it a rather fair metric [31]. HV can measure both convergence and diversity of a solution set in a sense, and larger values means better quality.
For calculating HV, the choice of reference point is a crucial issue. In our experiments, following the recommendation in [57] and [58], we set the reference point to be
Additionally, we use generational distance (GD) [60] and diversity comparison indicator (DCI) [61] as two assistant performance metrics. GD indicates how close the obtained solution set between the true PF, and can only reflect the convergence of an algorithm. For GD, smaller value is preferred. DCI is recently proposed for assessing the diversity of PF approximations in many-objective optimization.
C. MOEAs for Comparison
Eight MOEAs are considered to evaluate MOEA/D-DU and EFR-RR, four out of which are MOEA/D variants.
The first MOEA/D variant is a slightly modified version of MOEA/D-DE [40]. For a fair comparison, we replace the differential evolution (DE) operator in MOEA/D-DE with the recombination schemes in MOEA/D-DU. Another reason for this is that the results in [26] have indicated that DE operator seems not quite suitable for MaOPs. In our experimental studies, this MOEA/D version is referred as MOEA/D for short, which is the base algorithm of MOEA/D-DU.
The other three compared MOEA/D variants are MOEA/D-STM [46], MOEA/D-GR [49], and I-DBEA [48], since they are somewhat similar to MOEA/D-DU. Based on the same reasons mentioned above, we replace the DE operators in MOEA/D-STM and MOEA/D-GR with the recombination schemes in MOEA/D-DU.
The original EFR [42] is also involved in comparison because it is the predecessor of EFR-RR.
Moreover, two nondecomposition-based MOEAs are adopted for comparison, i.e., grid-based evolutionary algorithm (GrEA) [20] and shift-based density estimation (SDE) [25], both of which were proposed specially for many-objective optimization and compared favorably with several state-of-the-art many-objective optimizers. For SDE, the version SPEA2+SDE is used here as it achieves the best overall performance among the three considered versions (NSGA-II+SDE, SPEA2+SDE, and PESA-II+SDE) in [25].
All seven MOEAs mentioned above except MOEA/D-STM and I-DBEA do not incorporate any explicit normalization techniques in their original studies. MOEA/D-STM uses a naive normalization procedure, while I-DBEA uses a sophisticated one. The normalization is not the major contribution of this paper. So, for a fair comparison, we just remove the normalization procedures in MOEA/D-STM and I-DBEA. The seven MOEAs together with MOEA/D-DU and EFR-RR are compared on normalized test problems in Section VII-A. This practice is to eliminate the influence of differently scaled objective values and the normalization procedure on the performance of MOEAs and purely verify the effectiveness of the proposed strategies.
To investigate the performance on scaled test problems, we further incorporate the normalization procedure provided in Section IV-C into MOEA/D-DU and EFR-RR, and compare them in Section VII-B with NSGA-III [26], one of whose features is to include an ingenious normalization procedure to handle differently scaled objective values.
Except GrEA and SDE, all the other concerned MOEAs use the same approach [50] to generate structured weight vectors. MOEA/D, MOEA/D-STM, MOEA/D-GR, MOEA/D-DU, EFR, and EFR-RR all use the modified Tchebycheff function defined in (4).
All MOEAs including MOEA/D-DU and EFR-RR are implemented in the jMetal framework [62], and run on an Intel Core i7 2.9 GHz processor with 8 GB of RAM.
D. Experimental Settings
The experimental settings include general settings and parameter settings. The general settings are as follows.
Number of Runs and Termination Criterion: Each algorithm is run 30 times independently on each test instance, and the average metric values are recorded. The algorithm terminates after
function evaluations per run.$20\,000 \times m$ Significance Test: To test the difference for statistical significance in some cases, the Wilcoxon signed-rank test [63] at a 5% significance level is conducted on the metric values obtained by two competing algorithms.
As for parameter settings, we first list several common ones.
Population Size: The population size is the same with the number of weight vectors in all the algorithms except GrEA and SDE. Because of the generation mode of weight vectors, the population size for these algorithms cannot be arbitrary and is controlled by a parameter
($H$ ). As for GrEA and SDE, the population size can be set to any positive integer. But for ensuring a fair comparison, all the algorithms adopt the same population size for each problem instance. Table IV lists the population sizes used for varying number of objectives. To avoid only producing boundary weight vectors, the two layers of weight vectors [26] are adopted for problems having more than five objectives.$N = C_{H + m -1}^{m - 1}$ Parameter for Crossover and Mutation: All the algorithms use the SBX crossover and polynomial mutation to produce new solutions. The settings are shown in Table V. In EFR, EFR-RR, I-DBEA, and NSGA-III, a larger distribution index for crossover (
) is used according to [26], [42], and [48].$\eta _{c} =30$ Neighborhood Size
and Probability$T$ : In MOEA/D, MOEA/D-STM, MOEA/D-GR and MOEA/D-DU,$\delta $ is set to 20 and$T$ is set to 0.9.$\delta $ Parameter
in MOEA/D-DU and EFR-RR:$K$ is set to 5 for MOEA/D-DU and 2 for EFR-RR. In Section VI-A, the influence of$K$ will be investigated.$K$
MOEA/D, MOEA/D-GR, GrEA, and SDE have their specific parameters. In MOEA/D, the maximal number of solutions replaced by each child solution (
Analysis of the Performance of Enhanced Algorithms
In this section, the performance of enhanced algorithms is to be analyzed. First, the influence of parameter
A. Influence of Parameter $K$
Due to space limitation, Fig. 3 only shows the variation of HV values across all
exerts the similar impact on the performance of MOEA/D-DU and EFR-RR for each test instance.$K$ The most suitable setting of
not only depends on what problem to be solved but also depends on the number of objectives the problem has.$K$ For some problems, e.g., DTLZ4, the performance of MOEA/D-DU or EFR-RR is not very sensitive to the settings of
. Both of them can perform quite stably over a wide range of$K$ values.$K$ For some problems, e.g., WFG3 and WFG9, different setting of
may only lead to a slight variation of performance.$K$ For other problems, e.g., DTLZ7, a more careful setting of
should be conducted. In such cases, a slight variation of$K$ may result in a huge change of their performance.$K$
Examination of the influence of
Overall, although the performance of MOEA/D-DU and EFR-RR varies with different
B. Investigation of Convergence and Diversity
This section is devoted to investigate the convergence and diversity of the proposed algorithms using GD and DCI metrics, respectively. Here, we only consider normalized DTLZ1–DTLZ4 and WFG4–WFG9 problems. Because the PFs of these problems are all regular geometries, GD value can be easily determined analytically without sampling points on the PF. The PF shape of DTLZ1 is a hyperplane, while that of the other problems is a hypersphere.
We decide to compare our MOEA/D-DU and EFR-RR to their corresponding predecessors (MOEA/D and EFR) in the many-objective scenario. Table VII provides the average GD and DCI results. From Table VIII, both MOEA/D-DU and EFR-RR are generally better than their predecessors at maintaining the diversity, which achieves the better DCI results in almost all the concerned instances. For the convergence, MOEA/D-DU is still superior overall to MOEA/D and it obtains better GD results in most cases; EFR-RR is especially competitive to EFR and they perform better in terms of average GD with respect to different problem instances. It is worth noting that, compared with EFR, EFR-RR obtains much poorer GD and DCI results on 5-objective DTLZ3, which has a huge number of local PFs. The exact reasons for EFR-RRs underperformance on this instance is the subject of future investigation.
To describe the distribution of solutions in high-dimensional objective space more intuitively, Fig. 4 plots the final nondominated solutions of MOEA/D-DU, MOEA/D, EFR-RR, and EFR in a single run on 10-objective WFG4 instance by parallel coordinates. This particular run is associated with the result closest to the average HV value. As can be seen from Fig. 4, both MOEA/D-DU and EFR-RR can find a widely distributed solutions over
Final nondominated solution set of (a) MOEA/D-DU, (b) MOEA/D, (c) EFR-RR, and (d) EFR on the normalized 10-objective WFG4 instance, shown by parallel coordinates.
Based on the above comparisons, it can be concluded that MOEA/D-DU and EFR-RR can generally achieve better tradeoff between convergence and diversity than their corresponding predecessors.
Comparison With State-of-the-Art Algorithms
In this section, we compare the proposed MOEA/D-DU and EFR-RR with several state-of-the-art MOEAs according to the experimental design described in Section V.
A. Comparison on Normalized Test Problems
In this section, the algorithms without explicit normalization are compared on the first group of problems, i.e., normalized test problems. Table VIII shows the comparative results of nine algorithms in terms of HV metric on normalized DTLZ test problems, and Table IX on normalized WFG test problems. To have statistically sound conclusions, the significance test is conducted between MOEA/D-DU (EFR-RR) and each of seven other algorithms. In the two tables, the results significantly outperformed by MOEA/D-DU, EFR-RR, and both of them are marked in different symbols, respectively. Table X provides a summary of significance test of HV results between the proposed MOEA/D-DU (EFR-RR) and the other algorithms on all the considered 70 test instances. In this table,
From the above results, we can obtain the following observations for the proposed MOEA/D-DU and EFR-RR.
MOEA/D-DU is clearly superior to MOEA/D. Indeed, MOEA/D-DU significantly outperforms MOEA/D in 52 out of 70 instances, whereas MOEA/D only wins in eight instances and half of them are 2-objective ones. This strongly indicates that the proposed updating strategy in MOEA/D-DU is more effective than the original one in MOEA/D for solving MaOPs.
MOEA/D-DU also shows certain advantages over the other three MOEA/D variants, i.e., MOEA/D-STM, MOEA/D-GR and I-DBEA. Specifically, MOEA/D-DU significantly outperforms both MOEA/D-STM and MOEA/D-GR on vast majority of instances. Compared with I-DBEA, MOEA/D-DU is generally significantly better on 2, 5, and 8-objective instances, and it remains very competitive on 10-objective instances. But on 13-objective instances, I-DBEA generally shows higher performance. These results verify that the selection mechanism in MOEA/D-DU is comparable with or even better than the other alternative ones in state-of-the-art MOEA/D variants.
EFR-RR achieves an obvious improvement over EFR in many-objective scenario. Among 56 many-objective (
) instances, EFR-RR performs significantly better in 46 out of them, whereas EFR is significantly better in only 5 of them. For 2-objective instances, there exists no obvious performance difference between EFR-RR and EFR. These results clearly demonstrate the effectiveness of ranking restriction scheme used in EFR-RR on MaOPs.$m > 3$ MOEA/D-DU and EFR-RR are both particularly competitive to two nondecomposition-based MOEAs designed specially for many-objective optimization, i.e., GrEA and SDE. Each of the four algorithms performs best on some specific problems.
Based on the above extensive results, we further introduce the performance score [33] to quantify the overall performance of compared algorithms, which makes it easier to gain some insights into the behaviors of these algorithms. For a problem instance, suppose there are \begin{equation} P\left ({\text {Alg}_{i}}\right )=\sum _{\substack {j=1\\ j \neq i}}^{l}\delta _{i,j}. \end{equation}
Average performance score over all (a) normalized test problems for different number of objectives and (b) dimensions for different normalized test problems, namely DTLZ(Dx) and WFG(Wx). The smaller the score, the better the PF approximation in terms of HV metric. The values of proposed MOEA/D-DU and EFR-RR are connected by a solid line to easier assess the score.
From Fig. 5(a), both MOEA/D-DU and EFR-RR rank very well on MaOPs. For 2-objective problems, MOEA/D-DU is overall slightly better than MOEA/D, whereas EFR-RR is overall slightly worse than EFR. It is interesting to observe that, GrEA and I-DBEA, which show competitive performance on MaOPs, do not perform well on 2-objective problems. Conversely, MOEA/D, MOEA/D-STM, and MOEA/D-GR, which perform satisfactorily on 2-objective problems, scale poorly on MaOPs. Moreover, I-DBEA presents an interesting search behavior, i.e., it shows more competitiveness on problems having higher number of objectives.
From Fig. 5(b), EFR-RR achieves the best overall performance in 10 out of total 14 test problems, and it takes the second place on WFG9 problem. But EFR-RR does not show such excellent performance on WFG1 and WFG3. MOEA/D-DU performs overall best on WFG3, and it ranks second in eight instances, all following EFR-RR. However, MOEA/D-DU may struggle on DTLZ7 and WFG1, where it is only overall better than MOEA/D-GR.
Table XI presents the average performance score over all 70 problem instances for nine concerned algorithms, and the overall rank is given to each algorithm according to the score. The proposed EFR-RR and MOEA/D-DD are ranked first and second, respectively, followed by SDE and GrEA. But it is worth pointing out that this is just a comprehensive evaluation of algorithms on all the problem instances considered in this paper. Indeed, no algorithm can beat any of the other algorithms on all possible instances, and some algorithm is more suitable for solving certain kinds of problem, e.g., I-DBEA may show advantages over the others on problems with a very high number of objectives. In addition, the overall excellent performance of GrEA in our experiments is obtained by suitably setting div for each test instance. In this regard, GrEA takes advantage of the other compared algorithms.
Since GrEA and SDE are two competitive many-objective optimizers to MOEA/D-DU and EFR-RR, we further show the average CPU time over total
B. Comparison on Scaled Test Problems
In this section, we incorporate the normalization procedure presented in Section IV-C into the proposed MOEA/D-DU and EFR-RR, and compare them with NSGA-III on the second group of test problems, i.e., scaled test problems.
Table XIII shows the average HV results obtained by MOEA/D-DU, EFR-RR, and NSGA-III. From Table XIII, EFR-RR shows obvious advantage over NSGA-III. To be specific, EFR-RR significantly outperforms NSGA-III on 30 out of 40 instances, whereas NSGA-III is significantly better than EFR-RR only on five instances including 5- and 13-objective DTLZ1, 13-objective DTLZ2, and 5- and 13-objective WFG5. As for MOEA/D-DU, it is generally superior to NSGA-III on 2-, 5-, and 8-objective problems. On 10-objective problems, MOEA/D-DU and NSGA-III have equal shares, and either of them performs significantly better than the other on half of problems. Nevertheless, MOEA/D-DU does not perform as well as NSGA-III on 13-objective problems, where NSGA-III wins in 7 of them.
Fig. 6 plots the evolutionary trajectories of HV with generation numbers for MOEA/D-DU, EFR-RR and NSGA-III on 8-objective WFG9 instance. From Fig. 6, although all the three algorithms can stably increase HV with elapse of generation numbers, MOEA/D-DU and EFR-RR converge to the PF faster than NSGA-III and finally achieve higher HV values.
Evolutionary trajectories of HV over generations for MOEA/D-DU, EFR-RR, and NSGA-III on the 8-objective WFG9 instance.
From the above results, it can be concluded that either MOEA/D-DU or EFR-RR with an elaborate normalization procedure is at least competitive to or even better than NSGA-III for solving scaled problems.
C. Further Discussion
In this section, we would further discuss several issues concerning with the experimental studies.
The first issue is about the setting of population size in our experiments. Intuitively, exponentially more points are needed to represent a tradeoff of higher dimensional PF. Thus, a large population size is required for MOEAs to approximate the entire PF. However, just as mentioned in [26], it is difficult for a decision maker to comprehend and select a preferred solution in this scenario. Another drawback is that a large population size would make MOEAs computationally inefficient, and sometimes may be impractical. For example, if we use a population size of 100 in 2-objective case, then a population size of
The second issue is why MOEA/D-DU and EFR-RR do not exhibit clear superiority over their predecessors on 2-objective problems, although they have shown much better performance on many-objective ones. We suspect there are two possible reasons. First, just as mentioned above, we can only in fact use a population with the sparse distribution of solutions (and weight vectors) in a high-dimensional objective space, and a particular region may be associated with only one solution or one weight vector. So, if the aggregation function value is emphasized too much, the solution is very likely to deviate far from the corresponding region and thus to deteriorate the diversity preservation. In a 2-D objective space, just a normal population size, e.g., 100, can achieve a very dense distribution of solutions, so it is much less likely to fail to capture some regions of PF in this case even for MOEA/D or EFR. Second, there are more degrees of freedom in a higher dimensional objective space, so it is much more likely to produce the solutions that achieve a good aggregation value but far from the weight vector. Hence, the distance-based updating strategy in MOEA/D-DU and raking restriction scheme in EFR-RR would be more helpful to maintain the desired distribution of solutions in the evolutionary process.
The third issue is why two similar proposals to MOEA/D-DU, i.e., MOEA/D-STM and MOEA/D-GR, fail to show performance on par with MOEA/D-DU on MaOPs. For MOEA/D-STM, it uses a matching algorithm to coordinate the preferences of subproblems and solutions. Thus, it indeed implicitly gives equal importance to the two kinds of preferences. Nevertheless, it may be more appropriate to have a bias toward the preference of solutions in many-objective optimization since it is much more possible to lose track of some regions in high-dimensional objective space using sparsely distributed solutions. This is also experimentally verified by the parametric study of
The last issue is about the setting of parameter
Conclusion
In this paper, MOEA/D-DU and EFR-RR are proposed for many-objective optimization, which enhance two decomposition-based MOEAs, i.e., MOEA/D and EFR, respectively. The basic idea is to explicitly maintain the distribution of solutions during the evolutionary process by exploiting the perpendicular distance from a solution to a weight vector in the objective space. Specifically, in MOEA/D-DU, the updating procedure first determines
In the experimental studies, we have investigated the influence of a key parameter
In the future, we will study how to dynamically control the parameter