By Topic

On the influence of the instance structure on metaheuristic performances — application to a graph drawing problem

Sign In

Cookies must be enabled to login.After enabling cookies , please use refresh or reload or ctrl+f5 on the browser for the login options.

Formats Non-Member Member
$31 $13
Learn how you can qualify for the best price for this item!
Become an IEEE Member or Subscribe to
IEEE Xplore for exclusive pricing!
close button

puzzle piece

IEEE membership options for an individual and IEEE Xplore subscriptions for an organization offer the most affordable access to essential journal articles, conference papers, standards, eBooks, and eLearning courses.

Learn more about:

IEEE membership

IEEE Xplore subscriptions

2 Author(s)
Pinaud, B. ; Ecole Polytech. de l'' Univ. de Nantes, Nantes ; Kuntz, P.

Metaheuristics are now so common that for some classical hard combinatorial problems, there exist more than ten variants. Thus, the issue of comparing optimization methods is crucial. In this paper, we focus on one aspect of this question: the impact of the choice of the test instances on the metaheuristic performances and the possible link with the fitness landscape structure. We base our experimental framework on the arc crossing minimization problem for layered digraphs. We compare a hybridized genetic algorithm and a multistart descent which are among the best approaches to this problem. We worked on two instance families with various sizes and structural complexities: small graphs which are easy to draw on a standard size support, and large graphs specifically built for our experiments. We show that, for the smallest instances, there is no significant difference between methods whereas for graphs similar to those classically used nowadays in applications the genetic algorithm is better, and for the largest graphs (with a scaling factor up to 10300), the multistart descent is the best method. These results suggest that for "structured" fitness landscapes associated with real-life instances the GA exploits its implicit learning. On the other hand for very large landscapes with probably numerous local optima, only one exploration on a larger scale can be provided by local searches from a random starting point, cheap in computing effort.

Published in:

Evolutionary Computation, 2007. CEC 2007. IEEE Congress on

Date of Conference:

25-28 Sept. 2007