Enhancing Software Runtime with Reinforcement Learning-Driven Mutation Operator Selection in Genetic Improvement
Created by W.Langdon from
gp-bibliography.bib Revision:1.8237
- @InProceedings{bose:2025:GI,
-
author = "Damien Bose and Carol Hanna and Justyna Petke",
-
title = "Enhancing Software Runtime with Reinforcement
Learning-Driven Mutation Operator Selection in Genetic
Improvement",
-
booktitle = "14th International Workshop on Genetic Improvement
@ICSE 2025",
-
year = "2025",
-
editor = "Aymeric Blot and Vesna Nowack and
Penn {Faulkner Rainford} and Oliver Krauss",
-
address = "Ottawa",
-
month = "27 " # apr,
-
note = "forthcoming",
-
keywords = "genetic algorithms, genetic programming, Genetic
Improvement, Reinforcement learning",
-
URL = "
https://rps.ucl.ac.uk/viewobject.html?cid=1&id=2360068",
-
URL = "
https://gpbib.cs.ucl.ac.uk/gi2025/bose_2025_GI.pdf",
-
URL = "
https://geneticimprovementofsoftware.com/events/icse2025#accepted-papers",
-
size = "8 pages",
-
abstract = "Genetic Improvement employs heuristic search
algorithms to explore the search space of program
variants by modifying code using mutation operators.
This research focuses on operators that delete, insert
and replace source code statements. Traditionally, in
GI, an operator is chosen uniformly at random at each
search iteration. Reinforcement Learning to
intelligently guide the selection of these operators
specifically to improve program runtime. We propose to
integrate RL into the operator selection process. Four
Multi-Armed bandit RL algorithms (Epsilon Greedy, UCB,
Probability Matching, and Policy Gradient) were
integrated within a GI framework, and their efficacy
and efficiency were bench marked against the
traditional GI operator selection approach. These
RL-guided operator selection strategies have
demonstrated empirical superiority over the traditional
GI methods of randomly selecting a search operator,
with UCB emerging as the top-performing RL algorithm.
On average, the UCB-guided Hill Climbing search
algorithm produced variants that compiled and passed
all tests 44% of the time, while only 22% of the
variants produced by the traditional uniform random
selection strategies compiled and passed all tests.",
-
notes = "GI @ ICSE 2025, part of \cite{blot:2025:GI}
UCL ID: 10204649",
- }
Genetic Programming entries for
Damien Bose
Carol Hanna
Justyna Petke
Citations