z-logo
open-access-imgOpen Access
Backtracking Restarts for Deep Reinforcement Learning
Author(s) -
Zaid Marji,
John Licato
Publication year - 2021
Publication title -
proceedings of the ... international florida artificial intelligence research society conference
Language(s) - English
Resource type - Journals
eISSN - 2334-0762
pISSN - 2334-0754
DOI - 10.32473/flairs.v34i1.128557
Subject(s) - reinforcement learning , backtracking , computer science , markov decision process , intuition , artificial intelligence , markov chain , look ahead , markov process , machine learning , algorithm , mathematics , cognitive science , statistics , psychology
Manipulating the starting states of a Markov Decision Process to accelerate the learning of a deep reinforcement learning agent is an idea that has been proposed in several ways in the literature. Examples include starting from random states to improve exploration, taking random walks from desired goal states, and using performance-based metrics for starting states selection policy. In this paper, we explore the idea of exploiting the RL agent's trajectories generated during training for use as starting states. The main intuition behind this proposal is to focus the training of the RL agent to overcome its current weaknesses by practicing overcoming failure states by resetting the environment to a state in its recent past. We shall call the idea of starting from a fixed (or variable) number of steps back from recent terminal or failure states `backtracking restarts'. Our empirical findings show that this modification yields tangible speedups in the learning process.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here