Distance-based subset selection revisited

In this paper, we revisit the distance-based subset selection (DSS) algorithm in evolutionary multi-objective optimization. First, we show one drawback of the DSS algorithm, i.e., a uniformly distributed solution set cannot always be selected. Then, we show that this drawback can be overcome by maximizing the uniformity level of the selected solution set, which is defined by the minimum distance between two solutions in the solution set. Furthermore, we prove that the DSS algorithm is a greedy inclusion algorithm with respect to the maximization of the uniformity level. Based on this conclusion, we generalize DSS as a subset selection problem where the objective is to maximize the uniformity level of the subset. In addition to the greedy inclusion DSS algorithm, a greedy removal algorithm and an iterative algorithm are proposed for the generalized DSS problem. We also extend the Euclidean distance in the original DSS to other widely-used and user-defined distances. We conduct extensive experiments on solution sets over different types of Pareto fronts to compare the three DSS algorithms with different distances. Our results suggest the usefulness of the generalized DSS for selecting a uniform subset. The effect of using different distances on the selected subsets is also analyzed.

[1]  Hisao Ishibuchi,et al.  Benchmarking Multi- and Many-Objective Evolutionary Algorithms Under Two Optimization Scenarios , 2017, IEEE Access.

[2]  Carlos M. Fonseca,et al.  Greedy Hypervolume Subset Selection in Low Dimensions , 2016, Evolutionary Computation.

[3]  Nicola Beume,et al.  SMS-EMOA: Multiobjective selection based on dominated hypervolume , 2007, Eur. J. Oper. Res..

[4]  Hisao Ishibuchi,et al.  Selecting a small number of representative non-dominated solutions by a hypervolume-based solution selection approach , 2009, 2009 IEEE International Conference on Fuzzy Systems.

[5]  Carlos M. Fonseca,et al.  Hypervolume Subset Selection in Two Dimensions: Formulations and Algorithms , 2016, Evolutionary Computation.

[6]  Hisao Ishibuchi,et al.  A New Framework of Evolutionary Multi-Objective Algorithms with an Unbounded External Archive , 2020 .

[7]  Hisao Ishibuchi,et al.  Modified Distance Calculation in Generational Distance and Inverted Generational Distance , 2015, EMO.

[8]  Nicola Beume,et al.  An EMO Algorithm Using the Hypervolume Measure as Selection Criterion , 2005, EMO.

[9]  Lucas Bradstreet,et al.  Maximising Hypervolume for Selection in Multi-objective Evolutionary Algorithms , 2006, 2006 IEEE International Conference on Evolutionary Computation.

[10]  John E. Dennis,et al.  Normal-Boundary Intersection: A New Method for Generating the Pareto Surface in Nonlinear Multicriteria Optimization Problems , 1998, SIAM J. Optim..

[11]  Luís Paquete,et al.  Implicit enumeration strategies for the hypervolume subset selection problem , 2018, Comput. Oper. Res..

[12]  Hisao Ishibuchi,et al.  A Survey on the Hypervolume Indicator in Evolutionary Multiobjective Optimization , 2021, IEEE Transactions on Evolutionary Computation.

[13]  Frank Neumann,et al.  Maximizing Submodular Functions under Matroid Constraints by Evolutionary Algorithms , 2015, Evolutionary Computation.

[14]  Xin Yao,et al.  An Empirical Investigation of the Optimality and Monotonicity Properties of Multiobjective Archiving Methods , 2019, EMO.

[15]  Tapabrata Ray,et al.  Distance-Based Subset Selection for Benchmarking in Evolutionary Multi/Many-Objective Optimization , 2019, IEEE Transactions on Evolutionary Computation.

[16]  Serpil Sayin,et al.  Measuring the quality of discrete representations of efficient sets in multiple objective mathematical programming , 2000, Math. Program..