Cargando…
Entropy-Aware Model Initialization for Effective Exploration in Deep Reinforcement Learning
Effective exploration is one of the critical factors affecting performance in deep reinforcement learning. Agents acquire data to learn the optimal policy through exploration, and if it is not guaranteed, the data quality deteriorates, which leads to performance degradation. This study investigates...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9371101/ https://www.ncbi.nlm.nih.gov/pubmed/35957399 http://dx.doi.org/10.3390/s22155845 |
_version_ | 1784767031117086720 |
---|---|
author | Jang, Sooyoung Kim, Hyung-Il |
author_facet | Jang, Sooyoung Kim, Hyung-Il |
author_sort | Jang, Sooyoung |
collection | PubMed |
description | Effective exploration is one of the critical factors affecting performance in deep reinforcement learning. Agents acquire data to learn the optimal policy through exploration, and if it is not guaranteed, the data quality deteriorates, which leads to performance degradation. This study investigates the effect of initial entropy, which significantly influences exploration, especially in the early learning stage. The results of this study on tasks with discrete action space show that (1) low initial entropy increases the probability of learning failure, (2) the distributions of initial entropy for various tasks are biased towards low values that inhibit exploration, and (3) the initial entropy for discrete action space varies with both the initial weight and task, making it hard to control. We then devise a simple yet powerful learning strategy to deal with these limitations, namely, entropy-aware model initialization. The proposed algorithm aims to provide a model with high initial entropy to a deep reinforcement learning algorithm for effective exploration. Our experiments showed that the devised learning strategy significantly reduces learning failures and enhances performance, stability, and learning speed. |
format | Online Article Text |
id | pubmed-9371101 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-93711012022-08-12 Entropy-Aware Model Initialization for Effective Exploration in Deep Reinforcement Learning Jang, Sooyoung Kim, Hyung-Il Sensors (Basel) Article Effective exploration is one of the critical factors affecting performance in deep reinforcement learning. Agents acquire data to learn the optimal policy through exploration, and if it is not guaranteed, the data quality deteriorates, which leads to performance degradation. This study investigates the effect of initial entropy, which significantly influences exploration, especially in the early learning stage. The results of this study on tasks with discrete action space show that (1) low initial entropy increases the probability of learning failure, (2) the distributions of initial entropy for various tasks are biased towards low values that inhibit exploration, and (3) the initial entropy for discrete action space varies with both the initial weight and task, making it hard to control. We then devise a simple yet powerful learning strategy to deal with these limitations, namely, entropy-aware model initialization. The proposed algorithm aims to provide a model with high initial entropy to a deep reinforcement learning algorithm for effective exploration. Our experiments showed that the devised learning strategy significantly reduces learning failures and enhances performance, stability, and learning speed. MDPI 2022-08-04 /pmc/articles/PMC9371101/ /pubmed/35957399 http://dx.doi.org/10.3390/s22155845 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Jang, Sooyoung Kim, Hyung-Il Entropy-Aware Model Initialization for Effective Exploration in Deep Reinforcement Learning |
title | Entropy-Aware Model Initialization for Effective Exploration in Deep Reinforcement Learning |
title_full | Entropy-Aware Model Initialization for Effective Exploration in Deep Reinforcement Learning |
title_fullStr | Entropy-Aware Model Initialization for Effective Exploration in Deep Reinforcement Learning |
title_full_unstemmed | Entropy-Aware Model Initialization for Effective Exploration in Deep Reinforcement Learning |
title_short | Entropy-Aware Model Initialization for Effective Exploration in Deep Reinforcement Learning |
title_sort | entropy-aware model initialization for effective exploration in deep reinforcement learning |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9371101/ https://www.ncbi.nlm.nih.gov/pubmed/35957399 http://dx.doi.org/10.3390/s22155845 |
work_keys_str_mv | AT jangsooyoung entropyawaremodelinitializationforeffectiveexplorationindeepreinforcementlearning AT kimhyungil entropyawaremodelinitializationforeffectiveexplorationindeepreinforcementlearning |