Cargando…

Action-driven contrastive representation for reinforcement learning

In reinforcement learning, reward-driven feature learning directly from high-dimensional images faces two challenges: sample-efficiency for solving control tasks and generalization to unseen observations. In prior works, these issues have been addressed through learning representation from pixel inp...

Descripción completa

Detalles Bibliográficos
Autores principales: Kim, Minbeom, Rho, Kyeongha, Kim, Yong-duk, Jung, Kyomin
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Public Library of Science 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8932622/
https://www.ncbi.nlm.nih.gov/pubmed/35303031
http://dx.doi.org/10.1371/journal.pone.0265456
_version_ 1784671481840533504
author Kim, Minbeom
Rho, Kyeongha
Kim, Yong-duk
Jung, Kyomin
author_facet Kim, Minbeom
Rho, Kyeongha
Kim, Yong-duk
Jung, Kyomin
author_sort Kim, Minbeom
collection PubMed
description In reinforcement learning, reward-driven feature learning directly from high-dimensional images faces two challenges: sample-efficiency for solving control tasks and generalization to unseen observations. In prior works, these issues have been addressed through learning representation from pixel inputs. However, their representation faced the limitations of being vulnerable to the high diversity inherent in environments or not taking the characteristics for solving control tasks. To attenuate these phenomena, we propose the novel contrastive representation method, Action-Driven Auxiliary Task (ADAT), which forces a representation to concentrate on essential features for deciding actions and ignore control-irrelevant details. In the augmented state-action dictionary of ADAT, the agent learns representation to maximize agreement between observations sharing the same actions. The proposed method significantly outperforms model-free and model-based algorithms in the Atari and OpenAI ProcGen, widely used benchmarks for sample-efficiency and generalization.
format Online
Article
Text
id pubmed-8932622
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-89326222022-03-19 Action-driven contrastive representation for reinforcement learning Kim, Minbeom Rho, Kyeongha Kim, Yong-duk Jung, Kyomin PLoS One Research Article In reinforcement learning, reward-driven feature learning directly from high-dimensional images faces two challenges: sample-efficiency for solving control tasks and generalization to unseen observations. In prior works, these issues have been addressed through learning representation from pixel inputs. However, their representation faced the limitations of being vulnerable to the high diversity inherent in environments or not taking the characteristics for solving control tasks. To attenuate these phenomena, we propose the novel contrastive representation method, Action-Driven Auxiliary Task (ADAT), which forces a representation to concentrate on essential features for deciding actions and ignore control-irrelevant details. In the augmented state-action dictionary of ADAT, the agent learns representation to maximize agreement between observations sharing the same actions. The proposed method significantly outperforms model-free and model-based algorithms in the Atari and OpenAI ProcGen, widely used benchmarks for sample-efficiency and generalization. Public Library of Science 2022-03-18 /pmc/articles/PMC8932622/ /pubmed/35303031 http://dx.doi.org/10.1371/journal.pone.0265456 Text en © 2022 Kim et al https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
spellingShingle Research Article
Kim, Minbeom
Rho, Kyeongha
Kim, Yong-duk
Jung, Kyomin
Action-driven contrastive representation for reinforcement learning
title Action-driven contrastive representation for reinforcement learning
title_full Action-driven contrastive representation for reinforcement learning
title_fullStr Action-driven contrastive representation for reinforcement learning
title_full_unstemmed Action-driven contrastive representation for reinforcement learning
title_short Action-driven contrastive representation for reinforcement learning
title_sort action-driven contrastive representation for reinforcement learning
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8932622/
https://www.ncbi.nlm.nih.gov/pubmed/35303031
http://dx.doi.org/10.1371/journal.pone.0265456
work_keys_str_mv AT kimminbeom actiondrivencontrastiverepresentationforreinforcementlearning
AT rhokyeongha actiondrivencontrastiverepresentationforreinforcementlearning
AT kimyongduk actiondrivencontrastiverepresentationforreinforcementlearning
AT jungkyomin actiondrivencontrastiverepresentationforreinforcementlearning