Cargando…

Averaging Is Probably Not the Optimum Way of Aggregating Parameters in Federated Learning

Federated learning is a decentralized topology of deep learning, that trains a shared model through data distributed among each client (like mobile phones, wearable devices), in order to ensure data privacy by avoiding raw data exposed in data center (server). After each client computes a new model...

Descripción completa

Detalles Bibliográficos
Autores principales: Xiao, Peng, Cheng, Samuel, Stankovic, Vladimir, Vukobratovic, Dejan
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7516771/
https://www.ncbi.nlm.nih.gov/pubmed/33286088
http://dx.doi.org/10.3390/e22030314
_version_ 1783587078282412032
author Xiao, Peng
Cheng, Samuel
Stankovic, Vladimir
Vukobratovic, Dejan
author_facet Xiao, Peng
Cheng, Samuel
Stankovic, Vladimir
Vukobratovic, Dejan
author_sort Xiao, Peng
collection PubMed
description Federated learning is a decentralized topology of deep learning, that trains a shared model through data distributed among each client (like mobile phones, wearable devices), in order to ensure data privacy by avoiding raw data exposed in data center (server). After each client computes a new model parameter by stochastic gradient descent (SGD) based on their own local data, these locally-computed parameters will be aggregated to generate an updated global model. Many current state-of-the-art studies aggregate different client-computed parameters by averaging them, but none theoretically explains why averaging parameters is a good approach. In this paper, we treat each client computed parameter as a random vector because of the stochastic properties of SGD, and estimate mutual information between two client computed parameters at different training phases using two methods in two learning tasks. The results confirm the correlation between different clients and show an increasing trend of mutual information with training iteration. However, when we further compute the distance between client computed parameters, we find that parameters are getting more correlated while not getting closer. This phenomenon suggests that averaging parameters may not be the optimum way of aggregating trained parameters.
format Online
Article
Text
id pubmed-7516771
institution National Center for Biotechnology Information
language English
publishDate 2020
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-75167712020-11-09 Averaging Is Probably Not the Optimum Way of Aggregating Parameters in Federated Learning Xiao, Peng Cheng, Samuel Stankovic, Vladimir Vukobratovic, Dejan Entropy (Basel) Article Federated learning is a decentralized topology of deep learning, that trains a shared model through data distributed among each client (like mobile phones, wearable devices), in order to ensure data privacy by avoiding raw data exposed in data center (server). After each client computes a new model parameter by stochastic gradient descent (SGD) based on their own local data, these locally-computed parameters will be aggregated to generate an updated global model. Many current state-of-the-art studies aggregate different client-computed parameters by averaging them, but none theoretically explains why averaging parameters is a good approach. In this paper, we treat each client computed parameter as a random vector because of the stochastic properties of SGD, and estimate mutual information between two client computed parameters at different training phases using two methods in two learning tasks. The results confirm the correlation between different clients and show an increasing trend of mutual information with training iteration. However, when we further compute the distance between client computed parameters, we find that parameters are getting more correlated while not getting closer. This phenomenon suggests that averaging parameters may not be the optimum way of aggregating trained parameters. MDPI 2020-03-11 /pmc/articles/PMC7516771/ /pubmed/33286088 http://dx.doi.org/10.3390/e22030314 Text en © 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Xiao, Peng
Cheng, Samuel
Stankovic, Vladimir
Vukobratovic, Dejan
Averaging Is Probably Not the Optimum Way of Aggregating Parameters in Federated Learning
title Averaging Is Probably Not the Optimum Way of Aggregating Parameters in Federated Learning
title_full Averaging Is Probably Not the Optimum Way of Aggregating Parameters in Federated Learning
title_fullStr Averaging Is Probably Not the Optimum Way of Aggregating Parameters in Federated Learning
title_full_unstemmed Averaging Is Probably Not the Optimum Way of Aggregating Parameters in Federated Learning
title_short Averaging Is Probably Not the Optimum Way of Aggregating Parameters in Federated Learning
title_sort averaging is probably not the optimum way of aggregating parameters in federated learning
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7516771/
https://www.ncbi.nlm.nih.gov/pubmed/33286088
http://dx.doi.org/10.3390/e22030314
work_keys_str_mv AT xiaopeng averagingisprobablynottheoptimumwayofaggregatingparametersinfederatedlearning
AT chengsamuel averagingisprobablynottheoptimumwayofaggregatingparametersinfederatedlearning
AT stankovicvladimir averagingisprobablynottheoptimumwayofaggregatingparametersinfederatedlearning
AT vukobratovicdejan averagingisprobablynottheoptimumwayofaggregatingparametersinfederatedlearning