Evolutionary Computation, Volume 24, Issue 4, Page 573-575, Winter 2016.
These instructions give guidelines for preparing papers for this publication. Presents information for authors publishing in this journal.
Traditional multiobjective evolutionary algorithms face a great challenge when dealing with many objectives. This is due to a high proportion of nondominated solutions in the population and low selection pressure toward the Pareto front. In order to tackle this issue, a series of indicator-based algorithms have been proposed to guide the search process toward the Pareto front. However, a single indicator might be biased and lead the population to converge to a subregion of the Pareto front. In this paper, a multi-indicator-based algorithm is proposed for many-objective optimization problems. The proposed algorithm, namely stochastic ranking-based multi-indicator Algorithm (SRA), adopts the stochastic ranking technique to balance the search biases of different indicators. Empirical studies on a large number (39 in total) of problem instances from two well-defined benchmark sets with 5, 10, and 15 objectives demonstrate that SRA performs well in terms of inverted generational distance and hypervolume metrics when compared with state-of-the-art algorithms. Empirical studies also reveal that, in the case a problem requires the algorithm to have strong convergence ability, the performance of SRA can be further improved by incorporating a direction-based archive to store well-converged solutions and maintain diversity.
The generalized quadratic multiple knapsack problem (GQMKP) extends the classical quadratic multiple knapsack problem with setups and knapsack preference of the items. The GQMKP can accommodate a number of real-life applications and is computationally difficult. In this paper, we demonstrate the interest of the memetic search approach for approximating the GQMKP by presenting a highly effective memetic algorithm (denoted by MAGQMK). The algorithm combines a backbone-based crossover operator (to generate offspring solutions) and a multineighborhood simulated annealing procedure (to find high quality local optima). To prevent premature convergence of the search, MAGQMK employs a quality-and-distance (QD) pool updating strategy. Extensive experiments on two sets of 96 benchmarks show a remarkable performance of the proposed approach. In particular, it discovers improved best solutions in 53 and matches the best known solutions for 39 other cases. A case study on a pseudo real-life problem demonstrates the efficacy of the proposed approach in practical situations. Additional analyses show the important contribution of the novel general-exchange neighborhood, the backbone-based crossover operator as well as the QD pool updating rule to the performance of the proposed algorithm.
Describes the above-named upcoming special issue or section. May include topics to be covered or calls for papers.
The minimum weight-dominating set (MWDS) problem is NP-hard and has a lot of applications in the real world. Several metaheuristic methods have been developed for solving the problem effectively, but suffering from high CPU time on large-scale instances. In this paper, we design an effective hybrid memetic algorithm (HMA) for the MWDS problem. First, the MWDS problem is formulated as a constrained 0–1 programming problem and is converted to an equivalent unconstrained 0–1 problem using an adaptive penalty function. Then, we develop a memetic algorithm for the resulting problem, which contains a greedy randomized adaptive construction procedure, a tabu local search procedure, a crossover operator, a population-updating method, and a path-relinking procedure. These strategies make a good tradeoff between intensification and diversification. A number of experiments were carried out on three types of instances from the literature. Compared with existing algorithms, HMA is able to find high-quality solutions in much less CPU time. Specifically, HMA is at least six times faster than existing algorithms on the tested instances. With increasing instance size, the CPU time required by HMA increases much more slowly than required by existing algorithms.
A minimum Manhattan distance (MMD) approach to multiple criteria decision making in multiobjective optimization problems (MOPs) is proposed. The approach selects the final solution corresponding with a vector that has the MMD from a normalized ideal vector. This procedure is equivalent to the knee selection described by a divide and conquer approach that involves iterations of pairwise comparisons. Being able to systematically assign weighting coefficients to multiple criteria, the MMD approach is equivalent to a weighted-sum (WS) approach. Because of the equivalence, the MMD approach possesses rich geometric interpretations that are considered essential in the field of evolutionary computation. The MMD approach is elegant because all evaluations can be performed by efficient matrix calculations without iterations of comparisons. While the WS approach may encounter an indeterminate situation in which a few solutions yield almost the same WS, the MMD approach is able to determine the final solution discriminately. Since existing multiobjective evolutionary algorithms aim for a posteriori decision making, i.e., determining the final solution after a set of Pareto optimal solutions is available, the proposed MMD approach can be combined with them to form a powerful solution method of solving MOPs. Furthermore, the approach enables scalable definitions of the knee and knee solutions.
Module identification or community detection in complex networks has become increasingly important in many scientific fields because it provides insight into the relationship and interaction between network function and topology. In recent years, module identification algorithms based on stochastic optimization algorithms such as evolutionary algorithms have been demonstrated to be superior to other algorithms on small- to medium-scale networks. However, the scalability and resolution limit (RL) problems of these module identification algorithms have not been fully addressed, which impeded their application to real-world networks. This paper proposes a novel module identification algorithm called cooperative co-evolutionary module identification to address these two problems. The proposed algorithm employs a cooperative co-evolutionary framework to handle large-scale networks. We also incorporate a recursive partitioning scheme into the algorithm to effectively address the RL problem. The performance of our algorithm is evaluated on 12 benchmark complex networks. As a medical application, we apply our algorithm to identify disease modules that differentiate low- and high-grade glioma tumors to gain insights into the molecular mechanisms that underpin the progression of glioma. Experimental results show that the proposed algorithm has a very competitive performance compared with other state-of-the-art module identification algorithms.
Most existing work on evolutionary optimization assumes that there are analytic functions for evaluating the objectives and constraints. In the real world, however, the objective or constraint values of many optimization problems can be evaluated solely based on data and solving such optimization problems is often known as data-driven optimization. In this paper, we divide data-driven optimization problems into two categories, i.e., offline and online data-driven optimization, and discuss the main challenges involved therein. An evolutionary algorithm is then presented to optimize the design of a trauma system, which is a typical offline data-driven multiobjective optimization problem, where the objectives and constraints can be evaluated using incidents only. As each single function evaluation involves a large amount of patient data, we develop a multifidelity surrogate-management strategy to reduce the computation time of the evolutionary optimization. The main idea is to adaptively tune the approximation fidelity by clustering the original data into different numbers of clusters and a regression model is constructed to estimate the required minimum fidelity. Experimental results show that the proposed algorithm is able to save up to 90% of computation time without much sacrifice of the solution quality.