Cargando…

Improved Information-Theoretic Generalization Bounds for Distributed, Federated, and Iterative Learning †

We consider information-theoretic bounds on the expected generalization error for statistical learning problems in a network setting. In this setting, there are K nodes, each with its own independent dataset, and the models from the K nodes have to be aggregated into a final centralized model. We co...

Descripción completa

Detalles Bibliográficos
Autores principales: Barnes, Leighton Pate, Dytso, Alex, Poor, Harold Vincent
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9498125/
https://www.ncbi.nlm.nih.gov/pubmed/36141064
http://dx.doi.org/10.3390/e24091178
_version_ 1784794679377657856
author Barnes, Leighton Pate
Dytso, Alex
Poor, Harold Vincent
author_facet Barnes, Leighton Pate
Dytso, Alex
Poor, Harold Vincent
author_sort Barnes, Leighton Pate
collection PubMed
description We consider information-theoretic bounds on the expected generalization error for statistical learning problems in a network setting. In this setting, there are K nodes, each with its own independent dataset, and the models from the K nodes have to be aggregated into a final centralized model. We consider both simple averaging of the models as well as more complicated multi-round algorithms. We give upper bounds on the expected generalization error for a variety of problems, such as those with Bregman divergence or Lipschitz continuous losses, that demonstrate an improved dependence of [Formula: see text] on the number of nodes. These “per node” bounds are in terms of the mutual information between the training dataset and the trained weights at each node and are therefore useful in describing the generalization properties inherent to having communication or privacy constraints at each node.
format Online
Article
Text
id pubmed-9498125
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-94981252022-09-23 Improved Information-Theoretic Generalization Bounds for Distributed, Federated, and Iterative Learning † Barnes, Leighton Pate Dytso, Alex Poor, Harold Vincent Entropy (Basel) Article We consider information-theoretic bounds on the expected generalization error for statistical learning problems in a network setting. In this setting, there are K nodes, each with its own independent dataset, and the models from the K nodes have to be aggregated into a final centralized model. We consider both simple averaging of the models as well as more complicated multi-round algorithms. We give upper bounds on the expected generalization error for a variety of problems, such as those with Bregman divergence or Lipschitz continuous losses, that demonstrate an improved dependence of [Formula: see text] on the number of nodes. These “per node” bounds are in terms of the mutual information between the training dataset and the trained weights at each node and are therefore useful in describing the generalization properties inherent to having communication or privacy constraints at each node. MDPI 2022-08-24 /pmc/articles/PMC9498125/ /pubmed/36141064 http://dx.doi.org/10.3390/e24091178 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Barnes, Leighton Pate
Dytso, Alex
Poor, Harold Vincent
Improved Information-Theoretic Generalization Bounds for Distributed, Federated, and Iterative Learning †
title Improved Information-Theoretic Generalization Bounds for Distributed, Federated, and Iterative Learning †
title_full Improved Information-Theoretic Generalization Bounds for Distributed, Federated, and Iterative Learning †
title_fullStr Improved Information-Theoretic Generalization Bounds for Distributed, Federated, and Iterative Learning †
title_full_unstemmed Improved Information-Theoretic Generalization Bounds for Distributed, Federated, and Iterative Learning †
title_short Improved Information-Theoretic Generalization Bounds for Distributed, Federated, and Iterative Learning †
title_sort improved information-theoretic generalization bounds for distributed, federated, and iterative learning †
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9498125/
https://www.ncbi.nlm.nih.gov/pubmed/36141064
http://dx.doi.org/10.3390/e24091178
work_keys_str_mv AT barnesleightonpate improvedinformationtheoreticgeneralizationboundsfordistributedfederatedanditerativelearning
AT dytsoalex improvedinformationtheoreticgeneralizationboundsfordistributedfederatedanditerativelearning
AT poorharoldvincent improvedinformationtheoreticgeneralizationboundsfordistributedfederatedanditerativelearning