Cargando…

C2RL: Convolutional-Contrastive Learning for Reinforcement Learning Based on Self-Pretraining for Strong Augmentation

Reinforcement learning agents that have not been seen during training must be robust in test environments. However, the generalization problem is challenging to solve in reinforcement learning using high-dimensional images as the input. The addition of a self-supervised learning framework with data...

Descripción completa

Detalles Bibliográficos
Autores principales: Park, Sanghoon, Kim, Jihun, Jeong, Han-You, Kim, Tae-Kyoung, Yoo, Jinwoo
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10222541/
https://www.ncbi.nlm.nih.gov/pubmed/37430860
http://dx.doi.org/10.3390/s23104946
_version_ 1785049723713880064
author Park, Sanghoon
Kim, Jihun
Jeong, Han-You
Kim, Tae-Kyoung
Yoo, Jinwoo
author_facet Park, Sanghoon
Kim, Jihun
Jeong, Han-You
Kim, Tae-Kyoung
Yoo, Jinwoo
author_sort Park, Sanghoon
collection PubMed
description Reinforcement learning agents that have not been seen during training must be robust in test environments. However, the generalization problem is challenging to solve in reinforcement learning using high-dimensional images as the input. The addition of a self-supervised learning framework with data augmentation in the reinforcement learning architecture can promote generalization to a certain extent. However, excessively large changes in the input images may disturb reinforcement learning. Therefore, we propose a contrastive learning method that can help manage the trade-off relationship between the performance of reinforcement learning and auxiliary tasks against the data augmentation strength. In this framework, strong augmentation does not disturb reinforcement learning and instead maximizes the auxiliary effect for generalization. Results of experiments on the DeepMind Control suite demonstrate that the proposed method effectively uses strong data augmentation and achieves a higher generalization than the existing methods.
format Online
Article
Text
id pubmed-10222541
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-102225412023-05-28 C2RL: Convolutional-Contrastive Learning for Reinforcement Learning Based on Self-Pretraining for Strong Augmentation Park, Sanghoon Kim, Jihun Jeong, Han-You Kim, Tae-Kyoung Yoo, Jinwoo Sensors (Basel) Communication Reinforcement learning agents that have not been seen during training must be robust in test environments. However, the generalization problem is challenging to solve in reinforcement learning using high-dimensional images as the input. The addition of a self-supervised learning framework with data augmentation in the reinforcement learning architecture can promote generalization to a certain extent. However, excessively large changes in the input images may disturb reinforcement learning. Therefore, we propose a contrastive learning method that can help manage the trade-off relationship between the performance of reinforcement learning and auxiliary tasks against the data augmentation strength. In this framework, strong augmentation does not disturb reinforcement learning and instead maximizes the auxiliary effect for generalization. Results of experiments on the DeepMind Control suite demonstrate that the proposed method effectively uses strong data augmentation and achieves a higher generalization than the existing methods. MDPI 2023-05-21 /pmc/articles/PMC10222541/ /pubmed/37430860 http://dx.doi.org/10.3390/s23104946 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Communication
Park, Sanghoon
Kim, Jihun
Jeong, Han-You
Kim, Tae-Kyoung
Yoo, Jinwoo
C2RL: Convolutional-Contrastive Learning for Reinforcement Learning Based on Self-Pretraining for Strong Augmentation
title C2RL: Convolutional-Contrastive Learning for Reinforcement Learning Based on Self-Pretraining for Strong Augmentation
title_full C2RL: Convolutional-Contrastive Learning for Reinforcement Learning Based on Self-Pretraining for Strong Augmentation
title_fullStr C2RL: Convolutional-Contrastive Learning for Reinforcement Learning Based on Self-Pretraining for Strong Augmentation
title_full_unstemmed C2RL: Convolutional-Contrastive Learning for Reinforcement Learning Based on Self-Pretraining for Strong Augmentation
title_short C2RL: Convolutional-Contrastive Learning for Reinforcement Learning Based on Self-Pretraining for Strong Augmentation
title_sort c2rl: convolutional-contrastive learning for reinforcement learning based on self-pretraining for strong augmentation
topic Communication
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10222541/
https://www.ncbi.nlm.nih.gov/pubmed/37430860
http://dx.doi.org/10.3390/s23104946
work_keys_str_mv AT parksanghoon c2rlconvolutionalcontrastivelearningforreinforcementlearningbasedonselfpretrainingforstrongaugmentation
AT kimjihun c2rlconvolutionalcontrastivelearningforreinforcementlearningbasedonselfpretrainingforstrongaugmentation
AT jeonghanyou c2rlconvolutionalcontrastivelearningforreinforcementlearningbasedonselfpretrainingforstrongaugmentation
AT kimtaekyoung c2rlconvolutionalcontrastivelearningforreinforcementlearningbasedonselfpretrainingforstrongaugmentation
AT yoojinwoo c2rlconvolutionalcontrastivelearningforreinforcementlearningbasedonselfpretrainingforstrongaugmentation