Cargando…

Reinforcement Learning Algorithms for Autonomous Mission Accomplishment by Unmanned Aerial Vehicles: A Comparative View with DQN, SARSA and A2C

Unmanned aerial vehicles (UAV) can be controlled in diverse ways. One of the most common is through artificial intelligence (AI), which comprises different methods, such as reinforcement learning (RL). The article aims to provide a comparison of three RL algorithms—DQN as the benchmark, SARSA as a s...

Descripción completa

Detalles Bibliográficos
Autores principales: Jiménez, Gonzalo Aguilar, de la Escalera Hueso, Arturo, Gómez-Silva, Maria J.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10649256/
https://www.ncbi.nlm.nih.gov/pubmed/37960711
http://dx.doi.org/10.3390/s23219013
_version_ 1785135524894212096
author Jiménez, Gonzalo Aguilar
de la Escalera Hueso, Arturo
Gómez-Silva, Maria J.
author_facet Jiménez, Gonzalo Aguilar
de la Escalera Hueso, Arturo
Gómez-Silva, Maria J.
author_sort Jiménez, Gonzalo Aguilar
collection PubMed
description Unmanned aerial vehicles (UAV) can be controlled in diverse ways. One of the most common is through artificial intelligence (AI), which comprises different methods, such as reinforcement learning (RL). The article aims to provide a comparison of three RL algorithms—DQN as the benchmark, SARSA as a same-family algorithm, and A2C as a different-structure one—to address the problem of a UAV navigating from departure point A to endpoint B while avoiding obstacles and, simultaneously, using the least possible time and flying the shortest distance. Under fixed premises, this investigation provides the results of the performances obtained for this activity. A neighborhood environment was selected because it is likely one of the most common areas of use for commercial drones. Taking DQN as the benchmark and not having previous knowledge of the behavior of SARSA or A2C in the employed environment, the comparison outcomes showed that DQN was the only one achieving the target. At the same time, SARSA and A2C did not. However, a deeper analysis of the results led to the conclusion that a fine-tuning of A2C could overcome the performance of DQN under certain conditions, demonstrating a greater speed at maximum finding with a more straightforward structure.
format Online
Article
Text
id pubmed-10649256
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-106492562023-11-06 Reinforcement Learning Algorithms for Autonomous Mission Accomplishment by Unmanned Aerial Vehicles: A Comparative View with DQN, SARSA and A2C Jiménez, Gonzalo Aguilar de la Escalera Hueso, Arturo Gómez-Silva, Maria J. Sensors (Basel) Article Unmanned aerial vehicles (UAV) can be controlled in diverse ways. One of the most common is through artificial intelligence (AI), which comprises different methods, such as reinforcement learning (RL). The article aims to provide a comparison of three RL algorithms—DQN as the benchmark, SARSA as a same-family algorithm, and A2C as a different-structure one—to address the problem of a UAV navigating from departure point A to endpoint B while avoiding obstacles and, simultaneously, using the least possible time and flying the shortest distance. Under fixed premises, this investigation provides the results of the performances obtained for this activity. A neighborhood environment was selected because it is likely one of the most common areas of use for commercial drones. Taking DQN as the benchmark and not having previous knowledge of the behavior of SARSA or A2C in the employed environment, the comparison outcomes showed that DQN was the only one achieving the target. At the same time, SARSA and A2C did not. However, a deeper analysis of the results led to the conclusion that a fine-tuning of A2C could overcome the performance of DQN under certain conditions, demonstrating a greater speed at maximum finding with a more straightforward structure. MDPI 2023-11-06 /pmc/articles/PMC10649256/ /pubmed/37960711 http://dx.doi.org/10.3390/s23219013 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Jiménez, Gonzalo Aguilar
de la Escalera Hueso, Arturo
Gómez-Silva, Maria J.
Reinforcement Learning Algorithms for Autonomous Mission Accomplishment by Unmanned Aerial Vehicles: A Comparative View with DQN, SARSA and A2C
title Reinforcement Learning Algorithms for Autonomous Mission Accomplishment by Unmanned Aerial Vehicles: A Comparative View with DQN, SARSA and A2C
title_full Reinforcement Learning Algorithms for Autonomous Mission Accomplishment by Unmanned Aerial Vehicles: A Comparative View with DQN, SARSA and A2C
title_fullStr Reinforcement Learning Algorithms for Autonomous Mission Accomplishment by Unmanned Aerial Vehicles: A Comparative View with DQN, SARSA and A2C
title_full_unstemmed Reinforcement Learning Algorithms for Autonomous Mission Accomplishment by Unmanned Aerial Vehicles: A Comparative View with DQN, SARSA and A2C
title_short Reinforcement Learning Algorithms for Autonomous Mission Accomplishment by Unmanned Aerial Vehicles: A Comparative View with DQN, SARSA and A2C
title_sort reinforcement learning algorithms for autonomous mission accomplishment by unmanned aerial vehicles: a comparative view with dqn, sarsa and a2c
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10649256/
https://www.ncbi.nlm.nih.gov/pubmed/37960711
http://dx.doi.org/10.3390/s23219013
work_keys_str_mv AT jimenezgonzaloaguilar reinforcementlearningalgorithmsforautonomousmissionaccomplishmentbyunmannedaerialvehiclesacomparativeviewwithdqnsarsaanda2c
AT delaescalerahuesoarturo reinforcementlearningalgorithmsforautonomousmissionaccomplishmentbyunmannedaerialvehiclesacomparativeviewwithdqnsarsaanda2c
AT gomezsilvamariaj reinforcementlearningalgorithmsforautonomousmissionaccomplishmentbyunmannedaerialvehiclesacomparativeviewwithdqnsarsaanda2c