Cargando…

Multi-Agent Decision-Making Modes in Uncertain Interactive Traffic Scenarios via Graph Convolution-Based Deep Reinforcement Learning

As one of the main elements of reinforcement learning, the design of the reward function is often not given enough attention when reinforcement learning is used in concrete applications, which leads to unsatisfactory performances. In this study, a reward function matrix is proposed for training vari...

Descripción completa

Detalles Bibliográficos
Autores principales: Gao, Xin, Li, Xueyuan, Liu, Qi, Li, Zirui, Yang, Fan, Luan, Tian
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9230819/
https://www.ncbi.nlm.nih.gov/pubmed/35746364
http://dx.doi.org/10.3390/s22124586
_version_ 1784735163114061824
author Gao, Xin
Li, Xueyuan
Liu, Qi
Li, Zirui
Yang, Fan
Luan, Tian
author_facet Gao, Xin
Li, Xueyuan
Liu, Qi
Li, Zirui
Yang, Fan
Luan, Tian
author_sort Gao, Xin
collection PubMed
description As one of the main elements of reinforcement learning, the design of the reward function is often not given enough attention when reinforcement learning is used in concrete applications, which leads to unsatisfactory performances. In this study, a reward function matrix is proposed for training various decision-making modes with emphasis on decision-making styles and further emphasis on incentives and punishments. Additionally, we model a traffic scene via graph model to better represent the interaction between vehicles, and adopt the graph convolutional network (GCN) to extract the features of the graph structure to help the connected autonomous vehicles perform decision-making directly. Furthermore, we combine GCN with deep Q-learning and multi-step double deep Q-learning to train four decision-making modes, which are named the graph convolutional deep Q-network (GQN) and the multi-step double graph convolutional deep Q-network (MDGQN). In the simulation, the superiority of the reward function matrix is proved by comparing it with the baseline, and evaluation metrics are proposed to verify the performance differences among decision-making modes. Results show that the trained decision-making modes can satisfy various driving requirements, including task completion rate, safety requirements, comfort level, and completion efficiency, by adjusting the weight values in the reward function matrix. Finally, the decision-making modes trained by MDGQN had better performance in an uncertain highway exit scene than those trained by GQN.
format Online
Article
Text
id pubmed-9230819
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-92308192022-06-25 Multi-Agent Decision-Making Modes in Uncertain Interactive Traffic Scenarios via Graph Convolution-Based Deep Reinforcement Learning Gao, Xin Li, Xueyuan Liu, Qi Li, Zirui Yang, Fan Luan, Tian Sensors (Basel) Article As one of the main elements of reinforcement learning, the design of the reward function is often not given enough attention when reinforcement learning is used in concrete applications, which leads to unsatisfactory performances. In this study, a reward function matrix is proposed for training various decision-making modes with emphasis on decision-making styles and further emphasis on incentives and punishments. Additionally, we model a traffic scene via graph model to better represent the interaction between vehicles, and adopt the graph convolutional network (GCN) to extract the features of the graph structure to help the connected autonomous vehicles perform decision-making directly. Furthermore, we combine GCN with deep Q-learning and multi-step double deep Q-learning to train four decision-making modes, which are named the graph convolutional deep Q-network (GQN) and the multi-step double graph convolutional deep Q-network (MDGQN). In the simulation, the superiority of the reward function matrix is proved by comparing it with the baseline, and evaluation metrics are proposed to verify the performance differences among decision-making modes. Results show that the trained decision-making modes can satisfy various driving requirements, including task completion rate, safety requirements, comfort level, and completion efficiency, by adjusting the weight values in the reward function matrix. Finally, the decision-making modes trained by MDGQN had better performance in an uncertain highway exit scene than those trained by GQN. MDPI 2022-06-17 /pmc/articles/PMC9230819/ /pubmed/35746364 http://dx.doi.org/10.3390/s22124586 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Gao, Xin
Li, Xueyuan
Liu, Qi
Li, Zirui
Yang, Fan
Luan, Tian
Multi-Agent Decision-Making Modes in Uncertain Interactive Traffic Scenarios via Graph Convolution-Based Deep Reinforcement Learning
title Multi-Agent Decision-Making Modes in Uncertain Interactive Traffic Scenarios via Graph Convolution-Based Deep Reinforcement Learning
title_full Multi-Agent Decision-Making Modes in Uncertain Interactive Traffic Scenarios via Graph Convolution-Based Deep Reinforcement Learning
title_fullStr Multi-Agent Decision-Making Modes in Uncertain Interactive Traffic Scenarios via Graph Convolution-Based Deep Reinforcement Learning
title_full_unstemmed Multi-Agent Decision-Making Modes in Uncertain Interactive Traffic Scenarios via Graph Convolution-Based Deep Reinforcement Learning
title_short Multi-Agent Decision-Making Modes in Uncertain Interactive Traffic Scenarios via Graph Convolution-Based Deep Reinforcement Learning
title_sort multi-agent decision-making modes in uncertain interactive traffic scenarios via graph convolution-based deep reinforcement learning
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9230819/
https://www.ncbi.nlm.nih.gov/pubmed/35746364
http://dx.doi.org/10.3390/s22124586
work_keys_str_mv AT gaoxin multiagentdecisionmakingmodesinuncertaininteractivetrafficscenariosviagraphconvolutionbaseddeepreinforcementlearning
AT lixueyuan multiagentdecisionmakingmodesinuncertaininteractivetrafficscenariosviagraphconvolutionbaseddeepreinforcementlearning
AT liuqi multiagentdecisionmakingmodesinuncertaininteractivetrafficscenariosviagraphconvolutionbaseddeepreinforcementlearning
AT lizirui multiagentdecisionmakingmodesinuncertaininteractivetrafficscenariosviagraphconvolutionbaseddeepreinforcementlearning
AT yangfan multiagentdecisionmakingmodesinuncertaininteractivetrafficscenariosviagraphconvolutionbaseddeepreinforcementlearning
AT luantian multiagentdecisionmakingmodesinuncertaininteractivetrafficscenariosviagraphconvolutionbaseddeepreinforcementlearning