Cargando…

A Partition Based Gradient Compression Algorithm for Distributed Training in AIoT

Running Deep Neural Networks (DNNs) in distributed Internet of Things (IoT) nodes is a promising scheme to enhance the performance of IoT systems. However, due to the limited computing and communication resources of the IoT nodes, the communication efficiency of the distributed DNN training strategy...

Descripción completa

Detalles Bibliográficos
Autores principales: Guo, Bingjun, Liu, Yazhi, Zhang, Chunyang
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7999038/
https://www.ncbi.nlm.nih.gov/pubmed/33801972
http://dx.doi.org/10.3390/s21061943
_version_ 1783670690695610368
author Guo, Bingjun
Liu, Yazhi
Zhang, Chunyang
author_facet Guo, Bingjun
Liu, Yazhi
Zhang, Chunyang
author_sort Guo, Bingjun
collection PubMed
description Running Deep Neural Networks (DNNs) in distributed Internet of Things (IoT) nodes is a promising scheme to enhance the performance of IoT systems. However, due to the limited computing and communication resources of the IoT nodes, the communication efficiency of the distributed DNN training strategy is a problem demanding a prompt solution. In this paper, an adaptive compression strategy based on gradient partition is proposed to solve the problem of high communication overhead between nodes during the distributed training procedure. Firstly, a neural network is trained to predict the gradient distribution of its parameters. According to the distribution characteristics of the gradient, the gradient is divided into the key region and the sparse region. At the same time, combined with the information entropy of gradient distribution, a reasonable threshold is selected to filter the gradient value in the partition, and only the gradient value greater than the threshold is transmitted and updated, to reduce the traffic and improve the distributed training efficiency. The strategy uses gradient sparsity to achieve the maximum compression ratio of 37.1 times, which improves the training efficiency to a certain extent.
format Online
Article
Text
id pubmed-7999038
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-79990382021-03-28 A Partition Based Gradient Compression Algorithm for Distributed Training in AIoT Guo, Bingjun Liu, Yazhi Zhang, Chunyang Sensors (Basel) Article Running Deep Neural Networks (DNNs) in distributed Internet of Things (IoT) nodes is a promising scheme to enhance the performance of IoT systems. However, due to the limited computing and communication resources of the IoT nodes, the communication efficiency of the distributed DNN training strategy is a problem demanding a prompt solution. In this paper, an adaptive compression strategy based on gradient partition is proposed to solve the problem of high communication overhead between nodes during the distributed training procedure. Firstly, a neural network is trained to predict the gradient distribution of its parameters. According to the distribution characteristics of the gradient, the gradient is divided into the key region and the sparse region. At the same time, combined with the information entropy of gradient distribution, a reasonable threshold is selected to filter the gradient value in the partition, and only the gradient value greater than the threshold is transmitted and updated, to reduce the traffic and improve the distributed training efficiency. The strategy uses gradient sparsity to achieve the maximum compression ratio of 37.1 times, which improves the training efficiency to a certain extent. MDPI 2021-03-10 /pmc/articles/PMC7999038/ /pubmed/33801972 http://dx.doi.org/10.3390/s21061943 Text en © 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Guo, Bingjun
Liu, Yazhi
Zhang, Chunyang
A Partition Based Gradient Compression Algorithm for Distributed Training in AIoT
title A Partition Based Gradient Compression Algorithm for Distributed Training in AIoT
title_full A Partition Based Gradient Compression Algorithm for Distributed Training in AIoT
title_fullStr A Partition Based Gradient Compression Algorithm for Distributed Training in AIoT
title_full_unstemmed A Partition Based Gradient Compression Algorithm for Distributed Training in AIoT
title_short A Partition Based Gradient Compression Algorithm for Distributed Training in AIoT
title_sort partition based gradient compression algorithm for distributed training in aiot
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7999038/
https://www.ncbi.nlm.nih.gov/pubmed/33801972
http://dx.doi.org/10.3390/s21061943
work_keys_str_mv AT guobingjun apartitionbasedgradientcompressionalgorithmfordistributedtraininginaiot
AT liuyazhi apartitionbasedgradientcompressionalgorithmfordistributedtraininginaiot
AT zhangchunyang apartitionbasedgradientcompressionalgorithmfordistributedtraininginaiot
AT guobingjun partitionbasedgradientcompressionalgorithmfordistributedtraininginaiot
AT liuyazhi partitionbasedgradientcompressionalgorithmfordistributedtraininginaiot
AT zhangchunyang partitionbasedgradientcompressionalgorithmfordistributedtraininginaiot