Cargando…
Experience Replay Using Transition Sequences
Experience replay is one of the most commonly used approaches to improve the sample efficiency of reinforcement learning algorithms. In this work, we propose an approach to select and replay sequences of transitions in order to accelerate the learning of a reinforcement learning agent in an off-poli...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2018
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6022201/ https://www.ncbi.nlm.nih.gov/pubmed/29977200 http://dx.doi.org/10.3389/fnbot.2018.00032 |
_version_ | 1783335629588791296 |
---|---|
author | Karimpanal, Thommen George Bouffanais, Roland |
author_facet | Karimpanal, Thommen George Bouffanais, Roland |
author_sort | Karimpanal, Thommen George |
collection | PubMed |
description | Experience replay is one of the most commonly used approaches to improve the sample efficiency of reinforcement learning algorithms. In this work, we propose an approach to select and replay sequences of transitions in order to accelerate the learning of a reinforcement learning agent in an off-policy setting. In addition to selecting appropriate sequences, we also artificially construct transition sequences using information gathered from previous agent-environment interactions. These sequences, when replayed, allow value function information to trickle down to larger sections of the state/state-action space, thereby making the most of the agent's experience. We demonstrate our approach on modified versions of standard reinforcement learning tasks such as the mountain car and puddle world problems and empirically show that it enables faster, and more accurate learning of value functions as compared to other forms of experience replay. Further, we briefly discuss some of the possible extensions to this work, as well as applications and situations where this approach could be particularly useful. |
format | Online Article Text |
id | pubmed-6022201 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2018 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-60222012018-07-05 Experience Replay Using Transition Sequences Karimpanal, Thommen George Bouffanais, Roland Front Neurorobot Robotics and AI Experience replay is one of the most commonly used approaches to improve the sample efficiency of reinforcement learning algorithms. In this work, we propose an approach to select and replay sequences of transitions in order to accelerate the learning of a reinforcement learning agent in an off-policy setting. In addition to selecting appropriate sequences, we also artificially construct transition sequences using information gathered from previous agent-environment interactions. These sequences, when replayed, allow value function information to trickle down to larger sections of the state/state-action space, thereby making the most of the agent's experience. We demonstrate our approach on modified versions of standard reinforcement learning tasks such as the mountain car and puddle world problems and empirically show that it enables faster, and more accurate learning of value functions as compared to other forms of experience replay. Further, we briefly discuss some of the possible extensions to this work, as well as applications and situations where this approach could be particularly useful. Frontiers Media S.A. 2018-06-21 /pmc/articles/PMC6022201/ /pubmed/29977200 http://dx.doi.org/10.3389/fnbot.2018.00032 Text en Copyright © 2018 Karimpanal and Bouffanais. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Robotics and AI Karimpanal, Thommen George Bouffanais, Roland Experience Replay Using Transition Sequences |
title | Experience Replay Using Transition Sequences |
title_full | Experience Replay Using Transition Sequences |
title_fullStr | Experience Replay Using Transition Sequences |
title_full_unstemmed | Experience Replay Using Transition Sequences |
title_short | Experience Replay Using Transition Sequences |
title_sort | experience replay using transition sequences |
topic | Robotics and AI |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6022201/ https://www.ncbi.nlm.nih.gov/pubmed/29977200 http://dx.doi.org/10.3389/fnbot.2018.00032 |
work_keys_str_mv | AT karimpanalthommengeorge experiencereplayusingtransitionsequences AT bouffanaisroland experiencereplayusingtransitionsequences |