Cargando…

Accessibility of covariance information creates vulnerability in Federated Learning frameworks

MOTIVATION: Federated Learning (FL) is gaining traction in various fields as it enables integrative data analysis without sharing sensitive data, such as in healthcare. However, the risk of data leakage caused by malicious attacks must be considered. In this study, we introduce a novel attack algori...

Descripción completa

Detalles Bibliográficos
Autores principales: Huth, Manuel, Arruda, Jonas, Gusinow, Roy, Contento, Lorenzo, Tacconelli, Evelina, Hasenauer, Jan
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Oxford University Press 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10516515/
https://www.ncbi.nlm.nih.gov/pubmed/37647639
http://dx.doi.org/10.1093/bioinformatics/btad531
_version_ 1785109143027187712
author Huth, Manuel
Arruda, Jonas
Gusinow, Roy
Contento, Lorenzo
Tacconelli, Evelina
Hasenauer, Jan
author_facet Huth, Manuel
Arruda, Jonas
Gusinow, Roy
Contento, Lorenzo
Tacconelli, Evelina
Hasenauer, Jan
author_sort Huth, Manuel
collection PubMed
description MOTIVATION: Federated Learning (FL) is gaining traction in various fields as it enables integrative data analysis without sharing sensitive data, such as in healthcare. However, the risk of data leakage caused by malicious attacks must be considered. In this study, we introduce a novel attack algorithm that relies on being able to compute sample means, sample covariances, and construct known linearly independent vectors on the data owner side. RESULTS: We show that these basic functionalities, which are available in several established FL frameworks, are sufficient to reconstruct privacy-protected data. Additionally, the attack algorithm is robust to defense strategies that involve adding random noise. We demonstrate the limitations of existing frameworks and propose potential defense strategies analyzing the implications of using differential privacy. The novel insights presented in this study will aid in the improvement of FL frameworks. AVAILABILITY AND IMPLEMENTATION: The code examples are provided at GitHub (https://github.com/manuhuth/Data-Leakage-From-Covariances.git). The CNSIM1 dataset, which we used in the manuscript, is available within the DSData R package (https://github.com/datashield/DSData/tree/main/data).
format Online
Article
Text
id pubmed-10516515
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Oxford University Press
record_format MEDLINE/PubMed
spelling pubmed-105165152023-09-23 Accessibility of covariance information creates vulnerability in Federated Learning frameworks Huth, Manuel Arruda, Jonas Gusinow, Roy Contento, Lorenzo Tacconelli, Evelina Hasenauer, Jan Bioinformatics Original Paper MOTIVATION: Federated Learning (FL) is gaining traction in various fields as it enables integrative data analysis without sharing sensitive data, such as in healthcare. However, the risk of data leakage caused by malicious attacks must be considered. In this study, we introduce a novel attack algorithm that relies on being able to compute sample means, sample covariances, and construct known linearly independent vectors on the data owner side. RESULTS: We show that these basic functionalities, which are available in several established FL frameworks, are sufficient to reconstruct privacy-protected data. Additionally, the attack algorithm is robust to defense strategies that involve adding random noise. We demonstrate the limitations of existing frameworks and propose potential defense strategies analyzing the implications of using differential privacy. The novel insights presented in this study will aid in the improvement of FL frameworks. AVAILABILITY AND IMPLEMENTATION: The code examples are provided at GitHub (https://github.com/manuhuth/Data-Leakage-From-Covariances.git). The CNSIM1 dataset, which we used in the manuscript, is available within the DSData R package (https://github.com/datashield/DSData/tree/main/data). Oxford University Press 2023-08-30 /pmc/articles/PMC10516515/ /pubmed/37647639 http://dx.doi.org/10.1093/bioinformatics/btad531 Text en © The Author(s) 2023. Published by Oxford University Press. https://creativecommons.org/licenses/by/4.0/This is an Open Access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted reuse, distribution, and reproduction in any medium, provided the original work is properly cited.
spellingShingle Original Paper
Huth, Manuel
Arruda, Jonas
Gusinow, Roy
Contento, Lorenzo
Tacconelli, Evelina
Hasenauer, Jan
Accessibility of covariance information creates vulnerability in Federated Learning frameworks
title Accessibility of covariance information creates vulnerability in Federated Learning frameworks
title_full Accessibility of covariance information creates vulnerability in Federated Learning frameworks
title_fullStr Accessibility of covariance information creates vulnerability in Federated Learning frameworks
title_full_unstemmed Accessibility of covariance information creates vulnerability in Federated Learning frameworks
title_short Accessibility of covariance information creates vulnerability in Federated Learning frameworks
title_sort accessibility of covariance information creates vulnerability in federated learning frameworks
topic Original Paper
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10516515/
https://www.ncbi.nlm.nih.gov/pubmed/37647639
http://dx.doi.org/10.1093/bioinformatics/btad531
work_keys_str_mv AT huthmanuel accessibilityofcovarianceinformationcreatesvulnerabilityinfederatedlearningframeworks
AT arrudajonas accessibilityofcovarianceinformationcreatesvulnerabilityinfederatedlearningframeworks
AT gusinowroy accessibilityofcovarianceinformationcreatesvulnerabilityinfederatedlearningframeworks
AT contentolorenzo accessibilityofcovarianceinformationcreatesvulnerabilityinfederatedlearningframeworks
AT tacconellievelina accessibilityofcovarianceinformationcreatesvulnerabilityinfederatedlearningframeworks
AT hasenauerjan accessibilityofcovarianceinformationcreatesvulnerabilityinfederatedlearningframeworks