Cargando…

Traffic Signal Control Using Hybrid Action Space Deep Reinforcement Learning

Recent research works on intelligent traffic signal control (TSC) have been mainly focused on leveraging deep reinforcement learning (DRL) due to its proven capability and performance. DRL-based traffic signal control frameworks belong to either discrete or continuous controls. In discrete control,...

Descripción completa

Detalles Bibliográficos
Autores principales: Bouktif, Salah, Cheniki, Abderraouf, Ouni, Ali
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8037422/
https://www.ncbi.nlm.nih.gov/pubmed/33806123
http://dx.doi.org/10.3390/s21072302
_version_ 1783677139836469248
author Bouktif, Salah
Cheniki, Abderraouf
Ouni, Ali
author_facet Bouktif, Salah
Cheniki, Abderraouf
Ouni, Ali
author_sort Bouktif, Salah
collection PubMed
description Recent research works on intelligent traffic signal control (TSC) have been mainly focused on leveraging deep reinforcement learning (DRL) due to its proven capability and performance. DRL-based traffic signal control frameworks belong to either discrete or continuous controls. In discrete control, the DRL agent selects the appropriate traffic light phase from a finite set of phases. Whereas in continuous control approach, the agent decides the appropriate duration for each signal phase within a predetermined sequence of phases. Among the existing works, there are no prior approaches that propose a flexible framework combining both discrete and continuous DRL approaches in controlling traffic signal. Thus, our ultimate objective in this paper is to propose an approach capable of deciding simultaneously the proper phase and its associated duration. Our contribution resides in adapting a hybrid Deep Reinforcement Learning that considers at the same time discrete and continuous decisions. Precisely, we customize a Parameterized Deep Q-Networks (P-DQN) architecture that permits a hierarchical decision-making process that primarily decides the traffic light next phases and secondly specifies its the associated timing. The evaluation results of our approach using Simulation of Urban MObility (SUMO) shows its out-performance over the benchmarks. The proposed framework is able to reduce the average queue length of vehicles and the average travel time by 22.20% and 5.78%, respectively, over the alternative DRL-based TSC systems.
format Online
Article
Text
id pubmed-8037422
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-80374222021-04-12 Traffic Signal Control Using Hybrid Action Space Deep Reinforcement Learning Bouktif, Salah Cheniki, Abderraouf Ouni, Ali Sensors (Basel) Communication Recent research works on intelligent traffic signal control (TSC) have been mainly focused on leveraging deep reinforcement learning (DRL) due to its proven capability and performance. DRL-based traffic signal control frameworks belong to either discrete or continuous controls. In discrete control, the DRL agent selects the appropriate traffic light phase from a finite set of phases. Whereas in continuous control approach, the agent decides the appropriate duration for each signal phase within a predetermined sequence of phases. Among the existing works, there are no prior approaches that propose a flexible framework combining both discrete and continuous DRL approaches in controlling traffic signal. Thus, our ultimate objective in this paper is to propose an approach capable of deciding simultaneously the proper phase and its associated duration. Our contribution resides in adapting a hybrid Deep Reinforcement Learning that considers at the same time discrete and continuous decisions. Precisely, we customize a Parameterized Deep Q-Networks (P-DQN) architecture that permits a hierarchical decision-making process that primarily decides the traffic light next phases and secondly specifies its the associated timing. The evaluation results of our approach using Simulation of Urban MObility (SUMO) shows its out-performance over the benchmarks. The proposed framework is able to reduce the average queue length of vehicles and the average travel time by 22.20% and 5.78%, respectively, over the alternative DRL-based TSC systems. MDPI 2021-03-25 /pmc/articles/PMC8037422/ /pubmed/33806123 http://dx.doi.org/10.3390/s21072302 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) ).
spellingShingle Communication
Bouktif, Salah
Cheniki, Abderraouf
Ouni, Ali
Traffic Signal Control Using Hybrid Action Space Deep Reinforcement Learning
title Traffic Signal Control Using Hybrid Action Space Deep Reinforcement Learning
title_full Traffic Signal Control Using Hybrid Action Space Deep Reinforcement Learning
title_fullStr Traffic Signal Control Using Hybrid Action Space Deep Reinforcement Learning
title_full_unstemmed Traffic Signal Control Using Hybrid Action Space Deep Reinforcement Learning
title_short Traffic Signal Control Using Hybrid Action Space Deep Reinforcement Learning
title_sort traffic signal control using hybrid action space deep reinforcement learning
topic Communication
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8037422/
https://www.ncbi.nlm.nih.gov/pubmed/33806123
http://dx.doi.org/10.3390/s21072302
work_keys_str_mv AT bouktifsalah trafficsignalcontrolusinghybridactionspacedeepreinforcementlearning
AT chenikiabderraouf trafficsignalcontrolusinghybridactionspacedeepreinforcementlearning
AT ouniali trafficsignalcontrolusinghybridactionspacedeepreinforcementlearning