Cargando…

Fair detection of poisoning attacks in federated learning on non-i.i.d. data

Reconciling machine learning with individual privacy is one of the main motivations behind federated learning (FL), a decentralized machine learning technique that aggregates partial models trained by clients on their own private data to obtain a global deep learning model. Even if FL provides stron...

Descripción completa

Detalles Bibliográficos
Autores principales: Singh, Ashneet Khandpur, Blanco-Justicia, Alberto, Domingo-Ferrer, Josep
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Springer US 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9812008/
https://www.ncbi.nlm.nih.gov/pubmed/36619003
http://dx.doi.org/10.1007/s10618-022-00912-6
_version_ 1784863641156190208
author Singh, Ashneet Khandpur
Blanco-Justicia, Alberto
Domingo-Ferrer, Josep
author_facet Singh, Ashneet Khandpur
Blanco-Justicia, Alberto
Domingo-Ferrer, Josep
author_sort Singh, Ashneet Khandpur
collection PubMed
description Reconciling machine learning with individual privacy is one of the main motivations behind federated learning (FL), a decentralized machine learning technique that aggregates partial models trained by clients on their own private data to obtain a global deep learning model. Even if FL provides stronger privacy guarantees to the participating clients than centralized learning collecting the clients’ data in a central server, FL is vulnerable to some attacks whereby malicious clients submit bad updates in order to prevent the model from converging or, more subtly, to introduce artificial bias in the classification (poisoning). Poisoning detection techniques compute statistics on the updates to identify malicious clients. A downside of anti-poisoning techniques is that they might lead to discriminate minority groups whose data are significantly and legitimately different from those of the majority of clients. This would not only be unfair, but would yield poorer models that would fail to capture the knowledge in the training data, especially when data are not independent and identically distributed (non-i.i.d.). In this work, we strive to strike a balance between fighting poisoning and accommodating diversity to help learning fairer and less discriminatory federated learning models. In this way, we forestall the exclusion of diverse clients while still ensuring detection of poisoning attacks. Empirical work on three data sets shows that employing our approach to tell legitimate from malicious updates produces models that are more accurate than those obtained with state-of-the-art poisoning detection techniques. Additionally, we explore the impact of our proposal on the performance of models on non-i.i.d local training data.
format Online
Article
Text
id pubmed-9812008
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Springer US
record_format MEDLINE/PubMed
spelling pubmed-98120082023-01-04 Fair detection of poisoning attacks in federated learning on non-i.i.d. data Singh, Ashneet Khandpur Blanco-Justicia, Alberto Domingo-Ferrer, Josep Data Min Knowl Discov Article Reconciling machine learning with individual privacy is one of the main motivations behind federated learning (FL), a decentralized machine learning technique that aggregates partial models trained by clients on their own private data to obtain a global deep learning model. Even if FL provides stronger privacy guarantees to the participating clients than centralized learning collecting the clients’ data in a central server, FL is vulnerable to some attacks whereby malicious clients submit bad updates in order to prevent the model from converging or, more subtly, to introduce artificial bias in the classification (poisoning). Poisoning detection techniques compute statistics on the updates to identify malicious clients. A downside of anti-poisoning techniques is that they might lead to discriminate minority groups whose data are significantly and legitimately different from those of the majority of clients. This would not only be unfair, but would yield poorer models that would fail to capture the knowledge in the training data, especially when data are not independent and identically distributed (non-i.i.d.). In this work, we strive to strike a balance between fighting poisoning and accommodating diversity to help learning fairer and less discriminatory federated learning models. In this way, we forestall the exclusion of diverse clients while still ensuring detection of poisoning attacks. Empirical work on three data sets shows that employing our approach to tell legitimate from malicious updates produces models that are more accurate than those obtained with state-of-the-art poisoning detection techniques. Additionally, we explore the impact of our proposal on the performance of models on non-i.i.d local training data. Springer US 2023-01-04 /pmc/articles/PMC9812008/ /pubmed/36619003 http://dx.doi.org/10.1007/s10618-022-00912-6 Text en © The Author(s), under exclusive licence to Springer Science+Business Media LLC, part of Springer Nature 2023, Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law. This article is made available via the PMC Open Access Subset for unrestricted research re-use and secondary analysis in any form or by any means with acknowledgement of the original source. These permissions are granted for the duration of the World Health Organization (WHO) declaration of COVID-19 as a global pandemic.
spellingShingle Article
Singh, Ashneet Khandpur
Blanco-Justicia, Alberto
Domingo-Ferrer, Josep
Fair detection of poisoning attacks in federated learning on non-i.i.d. data
title Fair detection of poisoning attacks in federated learning on non-i.i.d. data
title_full Fair detection of poisoning attacks in federated learning on non-i.i.d. data
title_fullStr Fair detection of poisoning attacks in federated learning on non-i.i.d. data
title_full_unstemmed Fair detection of poisoning attacks in federated learning on non-i.i.d. data
title_short Fair detection of poisoning attacks in federated learning on non-i.i.d. data
title_sort fair detection of poisoning attacks in federated learning on non-i.i.d. data
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9812008/
https://www.ncbi.nlm.nih.gov/pubmed/36619003
http://dx.doi.org/10.1007/s10618-022-00912-6
work_keys_str_mv AT singhashneetkhandpur fairdetectionofpoisoningattacksinfederatedlearningonnoniiddata
AT blancojusticiaalberto fairdetectionofpoisoningattacksinfederatedlearningonnoniiddata
AT domingoferrerjosep fairdetectionofpoisoningattacksinfederatedlearningonnoniiddata