Cargando…
Performance Enhancement in Federated Learning by Reducing Class Imbalance of Non-IID Data
Due to the distributed data collection and learning in federated learnings, many clients conduct local training with non-independent and identically distributed (non-IID) datasets. Accordingly, the training from these datasets results in severe performance degradation. We propose an efficient algori...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9919903/ https://www.ncbi.nlm.nih.gov/pubmed/36772192 http://dx.doi.org/10.3390/s23031152 |
_version_ | 1784886938548830208 |
---|---|
author | Seol, Mihye Kim, Taejoon |
author_facet | Seol, Mihye Kim, Taejoon |
author_sort | Seol, Mihye |
collection | PubMed |
description | Due to the distributed data collection and learning in federated learnings, many clients conduct local training with non-independent and identically distributed (non-IID) datasets. Accordingly, the training from these datasets results in severe performance degradation. We propose an efficient algorithm for enhancing the performance of federated learning by overcoming the negative effects of non-IID datasets. First, the intra-client class imbalance is reduced by rendering the class distribution of clients close to Uniform distribution. Second, the clients to participate in federated learning are selected to make their integrated class distribution close to Uniform distribution for the purpose of mitigating the inter-client class imbalance, which represents the class distribution difference among clients. In addition, the amount of local training data for the selected clients is finely adjusted. Finally, in order to increase the efficiency of federated learning, the batch size and the learning rate of local training for the selected clients are dynamically controlled reflecting the effective size of the local dataset for each client. In the performance evaluation on CIFAR-10 and MNIST datasets, the proposed algorithm achieves 20% higher accuracy than existing federated learning algorithms. Moreover, in achieving this huge accuracy improvement, the proposed algorithm uses less computation and communication resources compared to existing algorithms in terms of the amount of data used and the number of clients joined in the training. |
format | Online Article Text |
id | pubmed-9919903 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-99199032023-02-12 Performance Enhancement in Federated Learning by Reducing Class Imbalance of Non-IID Data Seol, Mihye Kim, Taejoon Sensors (Basel) Article Due to the distributed data collection and learning in federated learnings, many clients conduct local training with non-independent and identically distributed (non-IID) datasets. Accordingly, the training from these datasets results in severe performance degradation. We propose an efficient algorithm for enhancing the performance of federated learning by overcoming the negative effects of non-IID datasets. First, the intra-client class imbalance is reduced by rendering the class distribution of clients close to Uniform distribution. Second, the clients to participate in federated learning are selected to make their integrated class distribution close to Uniform distribution for the purpose of mitigating the inter-client class imbalance, which represents the class distribution difference among clients. In addition, the amount of local training data for the selected clients is finely adjusted. Finally, in order to increase the efficiency of federated learning, the batch size and the learning rate of local training for the selected clients are dynamically controlled reflecting the effective size of the local dataset for each client. In the performance evaluation on CIFAR-10 and MNIST datasets, the proposed algorithm achieves 20% higher accuracy than existing federated learning algorithms. Moreover, in achieving this huge accuracy improvement, the proposed algorithm uses less computation and communication resources compared to existing algorithms in terms of the amount of data used and the number of clients joined in the training. MDPI 2023-01-19 /pmc/articles/PMC9919903/ /pubmed/36772192 http://dx.doi.org/10.3390/s23031152 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Seol, Mihye Kim, Taejoon Performance Enhancement in Federated Learning by Reducing Class Imbalance of Non-IID Data |
title | Performance Enhancement in Federated Learning by Reducing Class Imbalance of Non-IID Data |
title_full | Performance Enhancement in Federated Learning by Reducing Class Imbalance of Non-IID Data |
title_fullStr | Performance Enhancement in Federated Learning by Reducing Class Imbalance of Non-IID Data |
title_full_unstemmed | Performance Enhancement in Federated Learning by Reducing Class Imbalance of Non-IID Data |
title_short | Performance Enhancement in Federated Learning by Reducing Class Imbalance of Non-IID Data |
title_sort | performance enhancement in federated learning by reducing class imbalance of non-iid data |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9919903/ https://www.ncbi.nlm.nih.gov/pubmed/36772192 http://dx.doi.org/10.3390/s23031152 |
work_keys_str_mv | AT seolmihye performanceenhancementinfederatedlearningbyreducingclassimbalanceofnoniiddata AT kimtaejoon performanceenhancementinfederatedlearningbyreducingclassimbalanceofnoniiddata |