Cargando…
Adaptive Quadruped Balance Control for Dynamic Environments Using Maximum-Entropy Reinforcement Learning
External disturbance poses the primary threat to robot balance in dynamic environments. This paper provides a learning-based control architecture for quadrupedal self-balancing, which is adaptable to multiple unpredictable scenes of external continuous disturbance. Different from conventional method...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8434611/ https://www.ncbi.nlm.nih.gov/pubmed/34502796 http://dx.doi.org/10.3390/s21175907 |
_version_ | 1783751641017614336 |
---|---|
author | Sun, Haoran Fu, Tingting Ling, Yuanhuai He, Chaoming |
author_facet | Sun, Haoran Fu, Tingting Ling, Yuanhuai He, Chaoming |
author_sort | Sun, Haoran |
collection | PubMed |
description | External disturbance poses the primary threat to robot balance in dynamic environments. This paper provides a learning-based control architecture for quadrupedal self-balancing, which is adaptable to multiple unpredictable scenes of external continuous disturbance. Different from conventional methods which construct analytical models which explicitly reason the balancing process, our work utilized reinforcement learning and artificial neural network to avoid incomprehensible mathematical modeling. The control policy is composed of a neural network and a Tanh Gaussian policy, which implicitly establishes the fuzzy mapping from proprioceptive signals to action commands. During the training process, the maximum-entropy method (soft actor-critic algorithm) is employed to endow the policy with powerful exploration and generalization ability. The trained policy is validated in both simulations and realistic experiments with a customized quadruped robot. The results demonstrate that the policy can be easily transferred to the real world without elaborate configurations. Moreover, although this policy is trained in merely one specific vibration condition, it demonstrates robustness under conditions that were never encountered during training. |
format | Online Article Text |
id | pubmed-8434611 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-84346112021-09-12 Adaptive Quadruped Balance Control for Dynamic Environments Using Maximum-Entropy Reinforcement Learning Sun, Haoran Fu, Tingting Ling, Yuanhuai He, Chaoming Sensors (Basel) Article External disturbance poses the primary threat to robot balance in dynamic environments. This paper provides a learning-based control architecture for quadrupedal self-balancing, which is adaptable to multiple unpredictable scenes of external continuous disturbance. Different from conventional methods which construct analytical models which explicitly reason the balancing process, our work utilized reinforcement learning and artificial neural network to avoid incomprehensible mathematical modeling. The control policy is composed of a neural network and a Tanh Gaussian policy, which implicitly establishes the fuzzy mapping from proprioceptive signals to action commands. During the training process, the maximum-entropy method (soft actor-critic algorithm) is employed to endow the policy with powerful exploration and generalization ability. The trained policy is validated in both simulations and realistic experiments with a customized quadruped robot. The results demonstrate that the policy can be easily transferred to the real world without elaborate configurations. Moreover, although this policy is trained in merely one specific vibration condition, it demonstrates robustness under conditions that were never encountered during training. MDPI 2021-09-02 /pmc/articles/PMC8434611/ /pubmed/34502796 http://dx.doi.org/10.3390/s21175907 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Sun, Haoran Fu, Tingting Ling, Yuanhuai He, Chaoming Adaptive Quadruped Balance Control for Dynamic Environments Using Maximum-Entropy Reinforcement Learning |
title | Adaptive Quadruped Balance Control for Dynamic Environments Using Maximum-Entropy Reinforcement Learning |
title_full | Adaptive Quadruped Balance Control for Dynamic Environments Using Maximum-Entropy Reinforcement Learning |
title_fullStr | Adaptive Quadruped Balance Control for Dynamic Environments Using Maximum-Entropy Reinforcement Learning |
title_full_unstemmed | Adaptive Quadruped Balance Control for Dynamic Environments Using Maximum-Entropy Reinforcement Learning |
title_short | Adaptive Quadruped Balance Control for Dynamic Environments Using Maximum-Entropy Reinforcement Learning |
title_sort | adaptive quadruped balance control for dynamic environments using maximum-entropy reinforcement learning |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8434611/ https://www.ncbi.nlm.nih.gov/pubmed/34502796 http://dx.doi.org/10.3390/s21175907 |
work_keys_str_mv | AT sunhaoran adaptivequadrupedbalancecontrolfordynamicenvironmentsusingmaximumentropyreinforcementlearning AT futingting adaptivequadrupedbalancecontrolfordynamicenvironmentsusingmaximumentropyreinforcementlearning AT lingyuanhuai adaptivequadrupedbalancecontrolfordynamicenvironmentsusingmaximumentropyreinforcementlearning AT hechaoming adaptivequadrupedbalancecontrolfordynamicenvironmentsusingmaximumentropyreinforcementlearning |