Cargando…

Entropy-Aware Model Initialization for Effective Exploration in Deep Reinforcement Learning

Effective exploration is one of the critical factors affecting performance in deep reinforcement learning. Agents acquire data to learn the optimal policy through exploration, and if it is not guaranteed, the data quality deteriorates, which leads to performance degradation. This study investigates...

Descripción completa

Detalles Bibliográficos
Autores principales: Jang, Sooyoung, Kim, Hyung-Il
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9371101/
https://www.ncbi.nlm.nih.gov/pubmed/35957399
http://dx.doi.org/10.3390/s22155845
_version_ 1784767031117086720
author Jang, Sooyoung
Kim, Hyung-Il
author_facet Jang, Sooyoung
Kim, Hyung-Il
author_sort Jang, Sooyoung
collection PubMed
description Effective exploration is one of the critical factors affecting performance in deep reinforcement learning. Agents acquire data to learn the optimal policy through exploration, and if it is not guaranteed, the data quality deteriorates, which leads to performance degradation. This study investigates the effect of initial entropy, which significantly influences exploration, especially in the early learning stage. The results of this study on tasks with discrete action space show that (1) low initial entropy increases the probability of learning failure, (2) the distributions of initial entropy for various tasks are biased towards low values that inhibit exploration, and (3) the initial entropy for discrete action space varies with both the initial weight and task, making it hard to control. We then devise a simple yet powerful learning strategy to deal with these limitations, namely, entropy-aware model initialization. The proposed algorithm aims to provide a model with high initial entropy to a deep reinforcement learning algorithm for effective exploration. Our experiments showed that the devised learning strategy significantly reduces learning failures and enhances performance, stability, and learning speed.
format Online
Article
Text
id pubmed-9371101
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-93711012022-08-12 Entropy-Aware Model Initialization for Effective Exploration in Deep Reinforcement Learning Jang, Sooyoung Kim, Hyung-Il Sensors (Basel) Article Effective exploration is one of the critical factors affecting performance in deep reinforcement learning. Agents acquire data to learn the optimal policy through exploration, and if it is not guaranteed, the data quality deteriorates, which leads to performance degradation. This study investigates the effect of initial entropy, which significantly influences exploration, especially in the early learning stage. The results of this study on tasks with discrete action space show that (1) low initial entropy increases the probability of learning failure, (2) the distributions of initial entropy for various tasks are biased towards low values that inhibit exploration, and (3) the initial entropy for discrete action space varies with both the initial weight and task, making it hard to control. We then devise a simple yet powerful learning strategy to deal with these limitations, namely, entropy-aware model initialization. The proposed algorithm aims to provide a model with high initial entropy to a deep reinforcement learning algorithm for effective exploration. Our experiments showed that the devised learning strategy significantly reduces learning failures and enhances performance, stability, and learning speed. MDPI 2022-08-04 /pmc/articles/PMC9371101/ /pubmed/35957399 http://dx.doi.org/10.3390/s22155845 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Jang, Sooyoung
Kim, Hyung-Il
Entropy-Aware Model Initialization for Effective Exploration in Deep Reinforcement Learning
title Entropy-Aware Model Initialization for Effective Exploration in Deep Reinforcement Learning
title_full Entropy-Aware Model Initialization for Effective Exploration in Deep Reinforcement Learning
title_fullStr Entropy-Aware Model Initialization for Effective Exploration in Deep Reinforcement Learning
title_full_unstemmed Entropy-Aware Model Initialization for Effective Exploration in Deep Reinforcement Learning
title_short Entropy-Aware Model Initialization for Effective Exploration in Deep Reinforcement Learning
title_sort entropy-aware model initialization for effective exploration in deep reinforcement learning
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9371101/
https://www.ncbi.nlm.nih.gov/pubmed/35957399
http://dx.doi.org/10.3390/s22155845
work_keys_str_mv AT jangsooyoung entropyawaremodelinitializationforeffectiveexplorationindeepreinforcementlearning
AT kimhyungil entropyawaremodelinitializationforeffectiveexplorationindeepreinforcementlearning