Cargando…

Dual Dynamic Scheduling for Hierarchical QoS in Uplink-NOMA: A Reinforcement Learning Approach

The demand for bandwidth-intensive and delay-sensitive services is surging daily with the development of 5G technology, resulting in fierce competition for scarce radio resources. Power domain Nonorthogonal Multiple Access (NOMA) technologies can dramatically improve system capacity and spectrum eff...

Descripción completa

Detalles Bibliográficos
Autores principales: Li, Xiangjun, Cui, Qimei, Zhai, Jinli, Huang, Xueqing
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8271556/
https://www.ncbi.nlm.nih.gov/pubmed/34199075
http://dx.doi.org/10.3390/s21134404
_version_ 1783721030098878464
author Li, Xiangjun
Cui, Qimei
Zhai, Jinli
Huang, Xueqing
author_facet Li, Xiangjun
Cui, Qimei
Zhai, Jinli
Huang, Xueqing
author_sort Li, Xiangjun
collection PubMed
description The demand for bandwidth-intensive and delay-sensitive services is surging daily with the development of 5G technology, resulting in fierce competition for scarce radio resources. Power domain Nonorthogonal Multiple Access (NOMA) technologies can dramatically improve system capacity and spectrum efficiency. Unlike existing NOMA scheduling that mainly focuses on fairness, this paper proposes a power control solution for uplink hybrid OMA and PD-NOMA in dual dynamic environments: dynamic and imperfect channel information together with the random user-specific hierarchical quality of service (QoS). This paper models the power control problem as a nonconvex stochastic, which aims to maximize system energy efficiency while guaranteeing hierarchical user QoS requirements. Then, the problem is formulated as a partially observable Markov decision process (POMDP). Owing to the difficulty of modeling time-varying scenes, the urgency of fast convergency, the adaptability in a dynamic environment, and the continuity of the variables, a Deep Reinforcement Learning (DRL)-based method is proposed. This paper also transforms the hierarchical QoS constraint under the NOMA serial interference cancellation (SIC) scene to fit DRL. The simulation results verify the effectiveness and robustness of the proposed algorithm under a dual uncertain environment. As compared with the baseline Particle Swarm Optimization algorithm (PSO), the proposed DRL-based method has demonstrated satisfying performance.
format Online
Article
Text
id pubmed-8271556
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-82715562021-07-11 Dual Dynamic Scheduling for Hierarchical QoS in Uplink-NOMA: A Reinforcement Learning Approach Li, Xiangjun Cui, Qimei Zhai, Jinli Huang, Xueqing Sensors (Basel) Communication The demand for bandwidth-intensive and delay-sensitive services is surging daily with the development of 5G technology, resulting in fierce competition for scarce radio resources. Power domain Nonorthogonal Multiple Access (NOMA) technologies can dramatically improve system capacity and spectrum efficiency. Unlike existing NOMA scheduling that mainly focuses on fairness, this paper proposes a power control solution for uplink hybrid OMA and PD-NOMA in dual dynamic environments: dynamic and imperfect channel information together with the random user-specific hierarchical quality of service (QoS). This paper models the power control problem as a nonconvex stochastic, which aims to maximize system energy efficiency while guaranteeing hierarchical user QoS requirements. Then, the problem is formulated as a partially observable Markov decision process (POMDP). Owing to the difficulty of modeling time-varying scenes, the urgency of fast convergency, the adaptability in a dynamic environment, and the continuity of the variables, a Deep Reinforcement Learning (DRL)-based method is proposed. This paper also transforms the hierarchical QoS constraint under the NOMA serial interference cancellation (SIC) scene to fit DRL. The simulation results verify the effectiveness and robustness of the proposed algorithm under a dual uncertain environment. As compared with the baseline Particle Swarm Optimization algorithm (PSO), the proposed DRL-based method has demonstrated satisfying performance. MDPI 2021-06-27 /pmc/articles/PMC8271556/ /pubmed/34199075 http://dx.doi.org/10.3390/s21134404 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Communication
Li, Xiangjun
Cui, Qimei
Zhai, Jinli
Huang, Xueqing
Dual Dynamic Scheduling for Hierarchical QoS in Uplink-NOMA: A Reinforcement Learning Approach
title Dual Dynamic Scheduling for Hierarchical QoS in Uplink-NOMA: A Reinforcement Learning Approach
title_full Dual Dynamic Scheduling for Hierarchical QoS in Uplink-NOMA: A Reinforcement Learning Approach
title_fullStr Dual Dynamic Scheduling for Hierarchical QoS in Uplink-NOMA: A Reinforcement Learning Approach
title_full_unstemmed Dual Dynamic Scheduling for Hierarchical QoS in Uplink-NOMA: A Reinforcement Learning Approach
title_short Dual Dynamic Scheduling for Hierarchical QoS in Uplink-NOMA: A Reinforcement Learning Approach
title_sort dual dynamic scheduling for hierarchical qos in uplink-noma: a reinforcement learning approach
topic Communication
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8271556/
https://www.ncbi.nlm.nih.gov/pubmed/34199075
http://dx.doi.org/10.3390/s21134404
work_keys_str_mv AT lixiangjun dualdynamicschedulingforhierarchicalqosinuplinknomaareinforcementlearningapproach
AT cuiqimei dualdynamicschedulingforhierarchicalqosinuplinknomaareinforcementlearningapproach
AT zhaijinli dualdynamicschedulingforhierarchicalqosinuplinknomaareinforcementlearningapproach
AT huangxueqing dualdynamicschedulingforhierarchicalqosinuplinknomaareinforcementlearningapproach