Cargando…

Research on automatic pilot repetition generation method based on deep reinforcement learning

Using computers to replace pilot seats in air traffic control (ATC) simulators is an effective way to improve controller training efficiency and reduce training costs. To achieve this, we propose a deep reinforcement learning model, RoBERTa-RL (RoBERTa with Reinforcement Learning), for generating pi...

Descripción completa

Detalles Bibliográficos
Autores principales: Pan, Weijun, Jiang, Peiyuan, Li, Yukun, Wang, Zhuang, Huang, Junxiang
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10598579/
https://www.ncbi.nlm.nih.gov/pubmed/37885770
http://dx.doi.org/10.3389/fnbot.2023.1285831
_version_ 1785125585717035008
author Pan, Weijun
Jiang, Peiyuan
Li, Yukun
Wang, Zhuang
Huang, Junxiang
author_facet Pan, Weijun
Jiang, Peiyuan
Li, Yukun
Wang, Zhuang
Huang, Junxiang
author_sort Pan, Weijun
collection PubMed
description Using computers to replace pilot seats in air traffic control (ATC) simulators is an effective way to improve controller training efficiency and reduce training costs. To achieve this, we propose a deep reinforcement learning model, RoBERTa-RL (RoBERTa with Reinforcement Learning), for generating pilot repetitions. RoBERTa-RL is based on the pre-trained language model RoBERTa and is optimized through transfer learning and reinforcement learning. Transfer learning is used to address the issue of scarce data in the ATC domain, while reinforcement learning algorithms are employed to optimize the RoBERTa model and overcome the limitations in model generalization caused by transfer learning. We selected a real-world area control dataset as the target task training and testing dataset, and a tower control dataset generated based on civil aviation radio land-air communication rules as the test dataset for evaluating model generalization. In terms of the ROUGE evaluation metrics, RoBERTa-RL achieved significant results on the area control dataset with ROUGE-1, ROUGE-2, and ROUGE-L scores of 0.9962, 0.992, and 0.996, respectively. On the tower control dataset, the scores were 0.982, 0.954, and 0.982, respectively. To overcome the limitations of ROUGE in this field, we conducted a detailed evaluation of the proposed model architecture using keyword-based evaluation criteria for the generated repetition instructions. This evaluation criterion calculates various keyword-based metrics based on the segmented results of the repetition instruction text. In the keyword-based evaluation criteria, the constructed model achieved an overall accuracy of 98.8% on the area control dataset and 81.8% on the tower control dataset. In terms of generalization, RoBERTa-RL improved accuracy by 56% compared to the model before improvement and achieved a 47.5% improvement compared to various comparative models. These results indicate that employing reinforcement learning strategies to enhance deep learning algorithms can effectively mitigate the issue of poor generalization in text generation tasks, and this approach holds promise for future application in other related domains.
format Online
Article
Text
id pubmed-10598579
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-105985792023-10-26 Research on automatic pilot repetition generation method based on deep reinforcement learning Pan, Weijun Jiang, Peiyuan Li, Yukun Wang, Zhuang Huang, Junxiang Front Neurorobot Neuroscience Using computers to replace pilot seats in air traffic control (ATC) simulators is an effective way to improve controller training efficiency and reduce training costs. To achieve this, we propose a deep reinforcement learning model, RoBERTa-RL (RoBERTa with Reinforcement Learning), for generating pilot repetitions. RoBERTa-RL is based on the pre-trained language model RoBERTa and is optimized through transfer learning and reinforcement learning. Transfer learning is used to address the issue of scarce data in the ATC domain, while reinforcement learning algorithms are employed to optimize the RoBERTa model and overcome the limitations in model generalization caused by transfer learning. We selected a real-world area control dataset as the target task training and testing dataset, and a tower control dataset generated based on civil aviation radio land-air communication rules as the test dataset for evaluating model generalization. In terms of the ROUGE evaluation metrics, RoBERTa-RL achieved significant results on the area control dataset with ROUGE-1, ROUGE-2, and ROUGE-L scores of 0.9962, 0.992, and 0.996, respectively. On the tower control dataset, the scores were 0.982, 0.954, and 0.982, respectively. To overcome the limitations of ROUGE in this field, we conducted a detailed evaluation of the proposed model architecture using keyword-based evaluation criteria for the generated repetition instructions. This evaluation criterion calculates various keyword-based metrics based on the segmented results of the repetition instruction text. In the keyword-based evaluation criteria, the constructed model achieved an overall accuracy of 98.8% on the area control dataset and 81.8% on the tower control dataset. In terms of generalization, RoBERTa-RL improved accuracy by 56% compared to the model before improvement and achieved a 47.5% improvement compared to various comparative models. These results indicate that employing reinforcement learning strategies to enhance deep learning algorithms can effectively mitigate the issue of poor generalization in text generation tasks, and this approach holds promise for future application in other related domains. Frontiers Media S.A. 2023-10-11 /pmc/articles/PMC10598579/ /pubmed/37885770 http://dx.doi.org/10.3389/fnbot.2023.1285831 Text en Copyright © 2023 Pan, Jiang, Li, Wang and Huang. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Pan, Weijun
Jiang, Peiyuan
Li, Yukun
Wang, Zhuang
Huang, Junxiang
Research on automatic pilot repetition generation method based on deep reinforcement learning
title Research on automatic pilot repetition generation method based on deep reinforcement learning
title_full Research on automatic pilot repetition generation method based on deep reinforcement learning
title_fullStr Research on automatic pilot repetition generation method based on deep reinforcement learning
title_full_unstemmed Research on automatic pilot repetition generation method based on deep reinforcement learning
title_short Research on automatic pilot repetition generation method based on deep reinforcement learning
title_sort research on automatic pilot repetition generation method based on deep reinforcement learning
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10598579/
https://www.ncbi.nlm.nih.gov/pubmed/37885770
http://dx.doi.org/10.3389/fnbot.2023.1285831
work_keys_str_mv AT panweijun researchonautomaticpilotrepetitiongenerationmethodbasedondeepreinforcementlearning
AT jiangpeiyuan researchonautomaticpilotrepetitiongenerationmethodbasedondeepreinforcementlearning
AT liyukun researchonautomaticpilotrepetitiongenerationmethodbasedondeepreinforcementlearning
AT wangzhuang researchonautomaticpilotrepetitiongenerationmethodbasedondeepreinforcementlearning
AT huangjunxiang researchonautomaticpilotrepetitiongenerationmethodbasedondeepreinforcementlearning