@inproceedings{8e896a9e6fe94fb8bf0a788ccf7fd6da,
title = "Runtime analysis of (1 + 1) evolutionary algorithm controlled with Q-learning using greedy exploration strategy on OneMAX+ZEROMAX problem",
abstract = "There exist optimization problems with the target objective, which is to be optimized, and several extra objectives. The extra objectives may or may not be helpful in optimization process in terms of the number of objective evaluations necessary to reach an optimum of the target objective. OneMax+ZeroMax is a previously proposed benchmark optimization problem where the target objective is OneMax and a single extra objective is ZeroMax, which is equal to the number of zero bits in the bit vector. This is an example of a problem where extra objectives are not good, and objective selection methods should ignore the extra objectives. The EA+RL method is a method which selects objectives to be optimized by evolutionary algorithms (EA) using reinforcement learning (RL). Previously it was shown that it runs in Θ(N logN) on OneMax+ZeroMax when configured to use the randomized local search algorithm and the Q-learning algorithm with the greedy exploration strategy. We present the runtime analysis for the case when the (1 + 1)-EA algorithm is used. It is shown that the expected running time is at most 3.12eN log N.",
author = "Denis Antipov and Maxim Buzdalov and Benjamin Doerr",
year = "2015",
month = mar,
day = "15",
doi = "10.1007/978-3-319-16468-7_14",
language = "English",
isbn = "978-3-319-16467-0",
series = "Lecture Notes in Computer Science",
publisher = "Springer Nature",
pages = "160--172",
booktitle = "Evolutionary Computation in Combinatorial Optimization",
address = "Switzerland",
}