Asset Details
MbrlCatalogueTitleDetail
Do you wish to reserve the book?
The Univariate Marginal Distribution Algorithm Copes Well With Deception and Epistasis
by
Krejca, Martin S
, Doerr, Benjamin
in
Deception
/ Evolutionary algorithms
/ Fitness
/ Genetic algorithms
2020
Hey, we have placed the reservation for you!
By the way, why not check out events that you can attend while you pick your title.
You are currently in the queue to collect this book. You will be notified once it is your turn to collect the book.
Oops! Something went wrong.
Looks like we were not able to place the reservation. Kindly try again later.
Are you sure you want to remove the book from the shelf?
Oops! Something went wrong.
While trying to remove the title from your shelf something went wrong :( Kindly try again later!
Do you wish to request the book?
The Univariate Marginal Distribution Algorithm Copes Well With Deception and Epistasis
by
Krejca, Martin S
, Doerr, Benjamin
in
Deception
/ Evolutionary algorithms
/ Fitness
/ Genetic algorithms
2020
Please be aware that the book you have requested cannot be checked out. If you would like to checkout this book, you can reserve another copy
We have requested the book for you!
Your request is successful and it will be processed during the Library working hours. Please check the status of your request in My Requests.
Oops! Something went wrong.
Looks like we were not able to place your request. Kindly try again later.
The Univariate Marginal Distribution Algorithm Copes Well With Deception and Epistasis
Paper
The Univariate Marginal Distribution Algorithm Copes Well With Deception and Epistasis
2020
Request Book From Autostore
and Choose the Collection Method
Overview
In their recent work, Lehre and Nguyen (FOGA 2019) show that the univariate marginal distribution algorithm (UMDA) needs time exponential in the parent populations size to optimize the DeceptiveLeadingBlocks (DLB) problem. They conclude from this result that univariate EDAs have difficulties with deception and epistasis. In this work, we show that this negative finding is caused by an unfortunate choice of the parameters of the UMDA. When the population sizes are chosen large enough to prevent genetic drift, then the UMDA optimizes the DLB problem with high probability with at most \\(\\lambda(\\frac{n}{2} + 2 e \\ln n)\\) fitness evaluations. Since an offspring population size \\(\\lambda\\) of order \\(n \\log n\\) can prevent genetic drift, the UMDA can solve the DLB problem with \\(O(n^2 \\log n)\\) fitness evaluations. In contrast, for classic evolutionary algorithms no better run time guarantee than \\(O(n^3)\\) is known (which we prove to be tight for the \\({(1+1)}\\) EA), so our result rather suggests that the UMDA can cope well with deception and epistatis. From a broader perspective, our result shows that the UMDA can cope better with local optima than evolutionary algorithms; such a result was previously known only for the compact genetic algorithm. Together with the lower bound of Lehre and Nguyen, our result for the first time rigorously proves that running EDAs in the regime with genetic drift can lead to drastic performance losses.
Publisher
Cornell University Library, arXiv.org
Subject
This website uses cookies to ensure you get the best experience on our website.