Cargando…
Distributed DRL-Based Computation Offloading Scheme for Improving QoE in Edge Computing Environments
Various edge collaboration schemes that rely on reinforcement learning (RL) have been proposed to improve the quality of experience (QoE). Deep RL (DRL) maximizes cumulative rewards through large-scale exploration and exploitation. However, the existing DRL schemes do not consider the temporal state...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10144645/ https://www.ncbi.nlm.nih.gov/pubmed/37112505 http://dx.doi.org/10.3390/s23084166 |
_version_ | 1785034147788488704 |
---|---|
author | Park, Jinho Chung, Kwangsue |
author_facet | Park, Jinho Chung, Kwangsue |
author_sort | Park, Jinho |
collection | PubMed |
description | Various edge collaboration schemes that rely on reinforcement learning (RL) have been proposed to improve the quality of experience (QoE). Deep RL (DRL) maximizes cumulative rewards through large-scale exploration and exploitation. However, the existing DRL schemes do not consider the temporal states using a fully connected layer. Moreover, they learn the offloading policy regardless of the importance of experience. They also do not learn enough because of their limited experiences in distributed environments. To solve these problems, we proposed a distributed DRL-based computation offloading scheme for improving the QoE in edge computing environments. The proposed scheme selects the offloading target by modeling the task service time and load balance. We implemented three methods to improve the learning performance. Firstly, the DRL scheme used the least absolute shrinkage and selection operator (LASSO) regression and attention layer to consider the temporal states. Secondly, we learned the optimal policy based on the importance of experience using the TD error and loss of the critic network. Finally, we adaptively shared the experience between agents, based on the strategy gradient, to solve the data sparsity problem. The simulation results showed that the proposed scheme achieved lower variation and higher rewards than the existing schemes. |
format | Online Article Text |
id | pubmed-10144645 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-101446452023-04-29 Distributed DRL-Based Computation Offloading Scheme for Improving QoE in Edge Computing Environments Park, Jinho Chung, Kwangsue Sensors (Basel) Article Various edge collaboration schemes that rely on reinforcement learning (RL) have been proposed to improve the quality of experience (QoE). Deep RL (DRL) maximizes cumulative rewards through large-scale exploration and exploitation. However, the existing DRL schemes do not consider the temporal states using a fully connected layer. Moreover, they learn the offloading policy regardless of the importance of experience. They also do not learn enough because of their limited experiences in distributed environments. To solve these problems, we proposed a distributed DRL-based computation offloading scheme for improving the QoE in edge computing environments. The proposed scheme selects the offloading target by modeling the task service time and load balance. We implemented three methods to improve the learning performance. Firstly, the DRL scheme used the least absolute shrinkage and selection operator (LASSO) regression and attention layer to consider the temporal states. Secondly, we learned the optimal policy based on the importance of experience using the TD error and loss of the critic network. Finally, we adaptively shared the experience between agents, based on the strategy gradient, to solve the data sparsity problem. The simulation results showed that the proposed scheme achieved lower variation and higher rewards than the existing schemes. MDPI 2023-04-21 /pmc/articles/PMC10144645/ /pubmed/37112505 http://dx.doi.org/10.3390/s23084166 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Park, Jinho Chung, Kwangsue Distributed DRL-Based Computation Offloading Scheme for Improving QoE in Edge Computing Environments |
title | Distributed DRL-Based Computation Offloading Scheme for Improving QoE in Edge Computing Environments |
title_full | Distributed DRL-Based Computation Offloading Scheme for Improving QoE in Edge Computing Environments |
title_fullStr | Distributed DRL-Based Computation Offloading Scheme for Improving QoE in Edge Computing Environments |
title_full_unstemmed | Distributed DRL-Based Computation Offloading Scheme for Improving QoE in Edge Computing Environments |
title_short | Distributed DRL-Based Computation Offloading Scheme for Improving QoE in Edge Computing Environments |
title_sort | distributed drl-based computation offloading scheme for improving qoe in edge computing environments |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10144645/ https://www.ncbi.nlm.nih.gov/pubmed/37112505 http://dx.doi.org/10.3390/s23084166 |
work_keys_str_mv | AT parkjinho distributeddrlbasedcomputationoffloadingschemeforimprovingqoeinedgecomputingenvironments AT chungkwangsue distributeddrlbasedcomputationoffloadingschemeforimprovingqoeinedgecomputingenvironments |