Cargando…
A Turbo Q-Learning (TQL) for Energy Efficiency Optimization in Heterogeneous Networks
In order to maximize energy efficiency in heterogeneous networks (HetNets), a turbo Q-Learning (TQL) combined with multistage decision process and tabular Q-Learning is proposed to optimize the resource configuration. For the large dimensions of action space, the problem of energy efficiency optimiz...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2020
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7597261/ https://www.ncbi.nlm.nih.gov/pubmed/33286726 http://dx.doi.org/10.3390/e22090957 |
_version_ | 1783602305160970240 |
---|---|
author | Wang, Xiumin Li, Lei Li, Jun Li, Zhengquan |
author_facet | Wang, Xiumin Li, Lei Li, Jun Li, Zhengquan |
author_sort | Wang, Xiumin |
collection | PubMed |
description | In order to maximize energy efficiency in heterogeneous networks (HetNets), a turbo Q-Learning (TQL) combined with multistage decision process and tabular Q-Learning is proposed to optimize the resource configuration. For the large dimensions of action space, the problem of energy efficiency optimization is designed as a multistage decision process in this paper, according to the resource allocation of optimization objectives, the initial problem is divided into several subproblems which are solved by tabular Q-Learning, and the traditional exponential increasing size of action space is decomposed into linear increase. By iterating the solutions of subproblems, the initial problem is solved. The simple stability analysis of the algorithm is given in this paper. As to the large dimension of state space, we use a deep neural network (DNN) to classify states where the optimization policy of novel Q-Learning is set to label samples. Thus far, the dimensions of action and state space have been solved. The simulation results show that our approach is convergent, improves the convergence speed by 60% while maintaining almost the same energy efficiency and having the characteristics of system adjustment. |
format | Online Article Text |
id | pubmed-7597261 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2020 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-75972612020-11-09 A Turbo Q-Learning (TQL) for Energy Efficiency Optimization in Heterogeneous Networks Wang, Xiumin Li, Lei Li, Jun Li, Zhengquan Entropy (Basel) Article In order to maximize energy efficiency in heterogeneous networks (HetNets), a turbo Q-Learning (TQL) combined with multistage decision process and tabular Q-Learning is proposed to optimize the resource configuration. For the large dimensions of action space, the problem of energy efficiency optimization is designed as a multistage decision process in this paper, according to the resource allocation of optimization objectives, the initial problem is divided into several subproblems which are solved by tabular Q-Learning, and the traditional exponential increasing size of action space is decomposed into linear increase. By iterating the solutions of subproblems, the initial problem is solved. The simple stability analysis of the algorithm is given in this paper. As to the large dimension of state space, we use a deep neural network (DNN) to classify states where the optimization policy of novel Q-Learning is set to label samples. Thus far, the dimensions of action and state space have been solved. The simulation results show that our approach is convergent, improves the convergence speed by 60% while maintaining almost the same energy efficiency and having the characteristics of system adjustment. MDPI 2020-08-30 /pmc/articles/PMC7597261/ /pubmed/33286726 http://dx.doi.org/10.3390/e22090957 Text en © 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Wang, Xiumin Li, Lei Li, Jun Li, Zhengquan A Turbo Q-Learning (TQL) for Energy Efficiency Optimization in Heterogeneous Networks |
title | A Turbo Q-Learning (TQL) for Energy Efficiency Optimization in Heterogeneous Networks |
title_full | A Turbo Q-Learning (TQL) for Energy Efficiency Optimization in Heterogeneous Networks |
title_fullStr | A Turbo Q-Learning (TQL) for Energy Efficiency Optimization in Heterogeneous Networks |
title_full_unstemmed | A Turbo Q-Learning (TQL) for Energy Efficiency Optimization in Heterogeneous Networks |
title_short | A Turbo Q-Learning (TQL) for Energy Efficiency Optimization in Heterogeneous Networks |
title_sort | turbo q-learning (tql) for energy efficiency optimization in heterogeneous networks |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7597261/ https://www.ncbi.nlm.nih.gov/pubmed/33286726 http://dx.doi.org/10.3390/e22090957 |
work_keys_str_mv | AT wangxiumin aturboqlearningtqlforenergyefficiencyoptimizationinheterogeneousnetworks AT lilei aturboqlearningtqlforenergyefficiencyoptimizationinheterogeneousnetworks AT lijun aturboqlearningtqlforenergyefficiencyoptimizationinheterogeneousnetworks AT lizhengquan aturboqlearningtqlforenergyefficiencyoptimizationinheterogeneousnetworks AT wangxiumin turboqlearningtqlforenergyefficiencyoptimizationinheterogeneousnetworks AT lilei turboqlearningtqlforenergyefficiencyoptimizationinheterogeneousnetworks AT lijun turboqlearningtqlforenergyefficiencyoptimizationinheterogeneousnetworks AT lizhengquan turboqlearningtqlforenergyefficiencyoptimizationinheterogeneousnetworks |