Cargando…

D2D Mobile Relaying Meets NOMA—Part II: A Reinforcement Learning Perspective

Structureless communications such as Device-to-Device (D2D) relaying are undeniably of paramount importance to improving the performance of today’s mobile networks. Such a communication paradigm requires a certain level of intelligence at the device level, thereby allowing it to interact with the en...

Descripción completa

Detalles Bibliográficos
Autores principales: Driouech, Safaa, Sabir, Essaid, Ghogho, Mounir, Amhoud, El-Mehdi
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7961575/
https://www.ncbi.nlm.nih.gov/pubmed/33806302
http://dx.doi.org/10.3390/s21051755
_version_ 1783665290958077952
author Driouech, Safaa
Sabir, Essaid
Ghogho, Mounir
Amhoud, El-Mehdi
author_facet Driouech, Safaa
Sabir, Essaid
Ghogho, Mounir
Amhoud, El-Mehdi
author_sort Driouech, Safaa
collection PubMed
description Structureless communications such as Device-to-Device (D2D) relaying are undeniably of paramount importance to improving the performance of today’s mobile networks. Such a communication paradigm requires a certain level of intelligence at the device level, thereby allowing it to interact with the environment and make proper decisions. However, decentralizing decision-making may induce paradoxical outcomes, resulting in a drop in performance, which sustains the design of self-organizing yet efficient systems. We propose that each device decides either to directly connect to the eNodeB or get access via another device through a D2D link. In the first part of this article, we describe a biform game framework to analyze the proposed self-organized system’s performance, under pure and mixed strategies. We use two reinforcement learning (RL) algorithms, enabling devices to self-organize and learn their pure/mixed equilibrium strategies in a fully distributed fashion. Decentralized RL algorithms are shown to play an important role in allowing devices to be self-organized and reach satisfactory performance with incomplete information or even under uncertainties. We point out through a simulation the importance of D2D relaying and assess how our learning schemes perform under slow/fast channel fading.
format Online
Article
Text
id pubmed-7961575
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-79615752021-03-17 D2D Mobile Relaying Meets NOMA—Part II: A Reinforcement Learning Perspective Driouech, Safaa Sabir, Essaid Ghogho, Mounir Amhoud, El-Mehdi Sensors (Basel) Article Structureless communications such as Device-to-Device (D2D) relaying are undeniably of paramount importance to improving the performance of today’s mobile networks. Such a communication paradigm requires a certain level of intelligence at the device level, thereby allowing it to interact with the environment and make proper decisions. However, decentralizing decision-making may induce paradoxical outcomes, resulting in a drop in performance, which sustains the design of self-organizing yet efficient systems. We propose that each device decides either to directly connect to the eNodeB or get access via another device through a D2D link. In the first part of this article, we describe a biform game framework to analyze the proposed self-organized system’s performance, under pure and mixed strategies. We use two reinforcement learning (RL) algorithms, enabling devices to self-organize and learn their pure/mixed equilibrium strategies in a fully distributed fashion. Decentralized RL algorithms are shown to play an important role in allowing devices to be self-organized and reach satisfactory performance with incomplete information or even under uncertainties. We point out through a simulation the importance of D2D relaying and assess how our learning schemes perform under slow/fast channel fading. MDPI 2021-03-04 /pmc/articles/PMC7961575/ /pubmed/33806302 http://dx.doi.org/10.3390/s21051755 Text en © 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Driouech, Safaa
Sabir, Essaid
Ghogho, Mounir
Amhoud, El-Mehdi
D2D Mobile Relaying Meets NOMA—Part II: A Reinforcement Learning Perspective
title D2D Mobile Relaying Meets NOMA—Part II: A Reinforcement Learning Perspective
title_full D2D Mobile Relaying Meets NOMA—Part II: A Reinforcement Learning Perspective
title_fullStr D2D Mobile Relaying Meets NOMA—Part II: A Reinforcement Learning Perspective
title_full_unstemmed D2D Mobile Relaying Meets NOMA—Part II: A Reinforcement Learning Perspective
title_short D2D Mobile Relaying Meets NOMA—Part II: A Reinforcement Learning Perspective
title_sort d2d mobile relaying meets noma—part ii: a reinforcement learning perspective
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7961575/
https://www.ncbi.nlm.nih.gov/pubmed/33806302
http://dx.doi.org/10.3390/s21051755
work_keys_str_mv AT driouechsafaa d2dmobilerelayingmeetsnomapartiiareinforcementlearningperspective
AT sabiressaid d2dmobilerelayingmeetsnomapartiiareinforcementlearningperspective
AT ghoghomounir d2dmobilerelayingmeetsnomapartiiareinforcementlearningperspective
AT amhoudelmehdi d2dmobilerelayingmeetsnomapartiiareinforcementlearningperspective