联系方式

  • QQ:99515681
  • 邮箱:99515681@qq.com
  • 工作时间:8:00-23:00
  • 微信:codinghelp

您当前位置:首页 >> Java编程Java编程

日期:2021-07-12 11:17

720 IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. 22, NO. 5, OCTOBER 2018

Standard Steady State Genetic Algorithms Can

Hillclimb Faster Than Mutation-Only

Evolutionary Algorithms

Dogan Corus, Member, IEEE, and Pietro S. Oliveto , Senior Member, IEEE

Abstract—Explaining to what extent the real power of genetic

algorithms (GAs) lies in the ability of crossover to recombine indi-

viduals into higher quality solutions is an important problem in

evolutionary computation. In this paper we show how the inter-

play between mutation and crossover can make GAs hillclimb

faster than their mutation-only counterparts. We devise a Markov

chain framework that allows to rigorously prove an upper bound

on the runtime of standard steady state GAs to hillclimb the

ONEMAX function. The bound establishes that the steady-state

GAs are 25% faster than all standard bit mutation-only evolu-

tionary algorithms with static mutation rate up to lower order

terms for moderate population sizes. The analysis also suggests

that larger populations may be faster than populations of size 2.

We present a lower bound for a greedy (2 + 1) GA that matches

the upper bound for populations larger than 2, rigorously proving

that two individuals cannot outperform larger population sizes

under greedy selection and greedy crossover up to lower order

terms. In complementary experiments the best population size

is greater than 2 and the greedy GAs are faster than standard

ones, further suggesting that the derived lower bound also holds

for the standard steady state (2 + 1) GA.

Index Terms—Algorithms design and analysis, genetic algo-

rithms (GAs), Markov processes, stochastic processes.

I. INTRODUCTION

GENETIC algorithms (GAs) rely on a population of indi-viduals that simultaneously explore the search space.

The main distinguishing features of GAs from other random-

ized search heuristics is their use of a population and crossover

to generate new solutions. Rather than slightly modifying the

current best solution as in more traditional heuristics, the idea

behind GAs is that new solutions are generated by recom-

bining individuals of the current population (i.e., crossover).

Such individuals are selected to reproduce probabilistically

according to their fitness (i.e., reproduction). Occasionally,

Manuscript received November 15, 2016; revised April 13, 2017,

July 7, 2017, and August 3, 2017; accepted August 5, 2017. Date of publi-

cation September 26, 2017; date of current version September 28, 2018. This

work was supported by EPSRC under Grant EP/M004252/1. (Corresponding

author: Pietro S. Oliveto.)

The authors are with the Rigorous Research Team, Algorithms Group,

Department of Computer Science, University of Sheffield, Sheffield S1 4DP,

U.K. (e-mail: d.corus@sheffield.ac.uk; p.oliveto@sheffield.ac.uk).

This paper has supplementary downloadable multimedia material available

at http://ieeexplore.ieee.org provided by the authors.

Color versions of one or more of the figures in this paper are available

online at http://ieeexplore.ieee.org.

Digital Object Identifier 10.1109/TEVC.2017.2745715

random mutations may slightly modify the offspring produced

by crossover. The original motivation behind these mutations

is to avoid that some genetic material may be lost forever, thus

allowing to avoid premature convergence [16], [18]. For these

reasons the GA community traditionally regards crossover as

the main search operator while mutation is considered a “back-

ground operator” [1], [16], [19] or a “secondary mechanism

of genetic adaptation” [18].

Explaining when and why GAs are effective has proved to

be a nontrivial task. Schema theory and its resulting building

block hypothesis [18] were devised to explain such working

principles. However, these theories did not allow to rigor-

ously characterize the behavior and performance of GAs.

The hypothesis was disputed when a class of functions (i.e.,

Royal Road), thought to be ideal for GAs, was designed and

experiments revealed that the simple (1+1) EA was more

efficient [20], [27].

Runtime analysis approaches have provided rigorous proofs

that crossover may indeed speed up the evolutionary process

of GAs in ideal conditions (i.e., if sufficient diversity is avail-

able in the population). The JUMP function was introduced by

Jansen and Wegener [22] as a first example, where crossover

considerably improves the expected runtime compared to

mutation-only evolutionary algorithms (EAs). The proof

required an unrealistically small crossover probability to allow

mutation alone to create the necessary population diversity

for the crossover operator to then escape the local optimum.

Dang et al. [6] recently showed that the sufficient diversity,

and even faster upper bounds on the runtime for not too large

jump gaps, can be achieved also for realistic crossover proba-

bilities by using diversity mechanisms. Further examples that

show the effectiveness of crossover have been given for both

artificially constructed functions and standard combinatorial

optimization problems (see the next section for an overview).

Excellent hillclimbing performance of crossover-based GAs

has been also proved. Doerr et al. [8], [9] proposed a

(1+(λ, λ)) GA which optimizes the ONEMAX function in

(n


版权所有:留学生编程辅导网 2020 All Rights Reserved 联系方式:QQ:99515681 微信:codinghelp 电子信箱:99515681@qq.com
免责声明:本站部分内容从网络整理而来,只供参考!如有版权问题可联系本站删除。 站长地图

python代写
微信客服:codinghelp