Cargando…
A Graph Neural Network Based Decentralized Learning Scheme
As an emerging paradigm considering data privacy and transmission efficiency, decentralized learning aims to acquire a global model using the training data distributed over many user devices. It is a challenging problem since link loss, partial device participation, and non-independent and identical...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8839979/ https://www.ncbi.nlm.nih.gov/pubmed/35161776 http://dx.doi.org/10.3390/s22031030 |
_version_ | 1784650504583774208 |
---|---|
author | Gao, Huiguo Lee, Mengyuan Yu, Guanding Zhou, Zhaolin |
author_facet | Gao, Huiguo Lee, Mengyuan Yu, Guanding Zhou, Zhaolin |
author_sort | Gao, Huiguo |
collection | PubMed |
description | As an emerging paradigm considering data privacy and transmission efficiency, decentralized learning aims to acquire a global model using the training data distributed over many user devices. It is a challenging problem since link loss, partial device participation, and non-independent and identically distributed (non-iid) data distribution would all deteriorate the performance of decentralized learning algorithms. Existing work may restrict to linear models or show poor performance over non-iid data. Therefore, in this paper, we propose a decentralized learning scheme based on distributed parallel stochastic gradient descent (DPSGD) and graph neural network (GNN) to deal with the above challenges. Specifically, each user device participating in the learning task utilizes local training data to compute local stochastic gradients and updates its own local model. Then, each device utilizes the GNN model and exchanges the model parameters with its neighbors to reach the average of resultant global models. The iteration repeats until the algorithm converges. Extensive simulation results over both iid and non-iid data validate the algorithm’s convergence to near optimal results and robustness to both link loss and partial device participation. |
format | Online Article Text |
id | pubmed-8839979 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-88399792022-02-13 A Graph Neural Network Based Decentralized Learning Scheme Gao, Huiguo Lee, Mengyuan Yu, Guanding Zhou, Zhaolin Sensors (Basel) Article As an emerging paradigm considering data privacy and transmission efficiency, decentralized learning aims to acquire a global model using the training data distributed over many user devices. It is a challenging problem since link loss, partial device participation, and non-independent and identically distributed (non-iid) data distribution would all deteriorate the performance of decentralized learning algorithms. Existing work may restrict to linear models or show poor performance over non-iid data. Therefore, in this paper, we propose a decentralized learning scheme based on distributed parallel stochastic gradient descent (DPSGD) and graph neural network (GNN) to deal with the above challenges. Specifically, each user device participating in the learning task utilizes local training data to compute local stochastic gradients and updates its own local model. Then, each device utilizes the GNN model and exchanges the model parameters with its neighbors to reach the average of resultant global models. The iteration repeats until the algorithm converges. Extensive simulation results over both iid and non-iid data validate the algorithm’s convergence to near optimal results and robustness to both link loss and partial device participation. MDPI 2022-01-28 /pmc/articles/PMC8839979/ /pubmed/35161776 http://dx.doi.org/10.3390/s22031030 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Gao, Huiguo Lee, Mengyuan Yu, Guanding Zhou, Zhaolin A Graph Neural Network Based Decentralized Learning Scheme |
title | A Graph Neural Network Based Decentralized Learning Scheme |
title_full | A Graph Neural Network Based Decentralized Learning Scheme |
title_fullStr | A Graph Neural Network Based Decentralized Learning Scheme |
title_full_unstemmed | A Graph Neural Network Based Decentralized Learning Scheme |
title_short | A Graph Neural Network Based Decentralized Learning Scheme |
title_sort | graph neural network based decentralized learning scheme |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8839979/ https://www.ncbi.nlm.nih.gov/pubmed/35161776 http://dx.doi.org/10.3390/s22031030 |
work_keys_str_mv | AT gaohuiguo agraphneuralnetworkbaseddecentralizedlearningscheme AT leemengyuan agraphneuralnetworkbaseddecentralizedlearningscheme AT yuguanding agraphneuralnetworkbaseddecentralizedlearningscheme AT zhouzhaolin agraphneuralnetworkbaseddecentralizedlearningscheme AT gaohuiguo graphneuralnetworkbaseddecentralizedlearningscheme AT leemengyuan graphneuralnetworkbaseddecentralizedlearningscheme AT yuguanding graphneuralnetworkbaseddecentralizedlearningscheme AT zhouzhaolin graphneuralnetworkbaseddecentralizedlearningscheme |