Cargando…
Packet Flow Capacity Autonomous Operation Based on Reinforcement Learning
As the dynamicity of the traffic increases, the need for self-network operation becomes more evident. One of the solutions that might bring cost savings to network operators is the dynamic capacity management of large packet flows, especially in the context of packet over optical networks. Machine L...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8709052/ https://www.ncbi.nlm.nih.gov/pubmed/34960400 http://dx.doi.org/10.3390/s21248306 |
_version_ | 1784622838908452864 |
---|---|
author | Barzegar, Sima Ruiz, Marc Velasco, Luis |
author_facet | Barzegar, Sima Ruiz, Marc Velasco, Luis |
author_sort | Barzegar, Sima |
collection | PubMed |
description | As the dynamicity of the traffic increases, the need for self-network operation becomes more evident. One of the solutions that might bring cost savings to network operators is the dynamic capacity management of large packet flows, especially in the context of packet over optical networks. Machine Learning, particularly Reinforcement Learning, seems to be an enabler for autonomicity as a result of its inherent capacity to learn from experience. However, precisely because of that, RL methods might not be able to provide the required performance (e.g., delay, packet loss, and capacity overprovisioning) when managing the capacity of packet flows, until they learn the optimal policy. In view of that, we propose a management lifecycle with three phases: (i) a self-tuned threshold-based approach operating just after the packet flow is set up and until enough data on the traffic characteristics are available; (ii) an RL operation based on models pre-trained with a generic traffic profile; and (iii) an RL operation with models trained for real traffic. Exhaustive simulation results confirm the poor performance of RL algorithms until the optimal policy is learnt and when traffic characteristics change over time, which prevents deploying such methods in operators’ networks. In contrast, the proposed lifecycle outperforms benchmarking approaches, achieving noticeable performance from the beginning of operation while showing robustness against traffic changes. |
format | Online Article Text |
id | pubmed-8709052 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-87090522021-12-25 Packet Flow Capacity Autonomous Operation Based on Reinforcement Learning Barzegar, Sima Ruiz, Marc Velasco, Luis Sensors (Basel) Article As the dynamicity of the traffic increases, the need for self-network operation becomes more evident. One of the solutions that might bring cost savings to network operators is the dynamic capacity management of large packet flows, especially in the context of packet over optical networks. Machine Learning, particularly Reinforcement Learning, seems to be an enabler for autonomicity as a result of its inherent capacity to learn from experience. However, precisely because of that, RL methods might not be able to provide the required performance (e.g., delay, packet loss, and capacity overprovisioning) when managing the capacity of packet flows, until they learn the optimal policy. In view of that, we propose a management lifecycle with three phases: (i) a self-tuned threshold-based approach operating just after the packet flow is set up and until enough data on the traffic characteristics are available; (ii) an RL operation based on models pre-trained with a generic traffic profile; and (iii) an RL operation with models trained for real traffic. Exhaustive simulation results confirm the poor performance of RL algorithms until the optimal policy is learnt and when traffic characteristics change over time, which prevents deploying such methods in operators’ networks. In contrast, the proposed lifecycle outperforms benchmarking approaches, achieving noticeable performance from the beginning of operation while showing robustness against traffic changes. MDPI 2021-12-12 /pmc/articles/PMC8709052/ /pubmed/34960400 http://dx.doi.org/10.3390/s21248306 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Barzegar, Sima Ruiz, Marc Velasco, Luis Packet Flow Capacity Autonomous Operation Based on Reinforcement Learning |
title | Packet Flow Capacity Autonomous Operation Based on Reinforcement Learning |
title_full | Packet Flow Capacity Autonomous Operation Based on Reinforcement Learning |
title_fullStr | Packet Flow Capacity Autonomous Operation Based on Reinforcement Learning |
title_full_unstemmed | Packet Flow Capacity Autonomous Operation Based on Reinforcement Learning |
title_short | Packet Flow Capacity Autonomous Operation Based on Reinforcement Learning |
title_sort | packet flow capacity autonomous operation based on reinforcement learning |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8709052/ https://www.ncbi.nlm.nih.gov/pubmed/34960400 http://dx.doi.org/10.3390/s21248306 |
work_keys_str_mv | AT barzegarsima packetflowcapacityautonomousoperationbasedonreinforcementlearning AT ruizmarc packetflowcapacityautonomousoperationbasedonreinforcementlearning AT velascoluis packetflowcapacityautonomousoperationbasedonreinforcementlearning |