Cargando…
Accessibility of covariance information creates vulnerability in Federated Learning frameworks
MOTIVATION: Federated Learning (FL) is gaining traction in various fields as it enables integrative data analysis without sharing sensitive data, such as in healthcare. However, the risk of data leakage caused by malicious attacks must be considered. In this study, we introduce a novel attack algori...
Autores principales: | , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Oxford University Press
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10516515/ https://www.ncbi.nlm.nih.gov/pubmed/37647639 http://dx.doi.org/10.1093/bioinformatics/btad531 |
_version_ | 1785109143027187712 |
---|---|
author | Huth, Manuel Arruda, Jonas Gusinow, Roy Contento, Lorenzo Tacconelli, Evelina Hasenauer, Jan |
author_facet | Huth, Manuel Arruda, Jonas Gusinow, Roy Contento, Lorenzo Tacconelli, Evelina Hasenauer, Jan |
author_sort | Huth, Manuel |
collection | PubMed |
description | MOTIVATION: Federated Learning (FL) is gaining traction in various fields as it enables integrative data analysis without sharing sensitive data, such as in healthcare. However, the risk of data leakage caused by malicious attacks must be considered. In this study, we introduce a novel attack algorithm that relies on being able to compute sample means, sample covariances, and construct known linearly independent vectors on the data owner side. RESULTS: We show that these basic functionalities, which are available in several established FL frameworks, are sufficient to reconstruct privacy-protected data. Additionally, the attack algorithm is robust to defense strategies that involve adding random noise. We demonstrate the limitations of existing frameworks and propose potential defense strategies analyzing the implications of using differential privacy. The novel insights presented in this study will aid in the improvement of FL frameworks. AVAILABILITY AND IMPLEMENTATION: The code examples are provided at GitHub (https://github.com/manuhuth/Data-Leakage-From-Covariances.git). The CNSIM1 dataset, which we used in the manuscript, is available within the DSData R package (https://github.com/datashield/DSData/tree/main/data). |
format | Online Article Text |
id | pubmed-10516515 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Oxford University Press |
record_format | MEDLINE/PubMed |
spelling | pubmed-105165152023-09-23 Accessibility of covariance information creates vulnerability in Federated Learning frameworks Huth, Manuel Arruda, Jonas Gusinow, Roy Contento, Lorenzo Tacconelli, Evelina Hasenauer, Jan Bioinformatics Original Paper MOTIVATION: Federated Learning (FL) is gaining traction in various fields as it enables integrative data analysis without sharing sensitive data, such as in healthcare. However, the risk of data leakage caused by malicious attacks must be considered. In this study, we introduce a novel attack algorithm that relies on being able to compute sample means, sample covariances, and construct known linearly independent vectors on the data owner side. RESULTS: We show that these basic functionalities, which are available in several established FL frameworks, are sufficient to reconstruct privacy-protected data. Additionally, the attack algorithm is robust to defense strategies that involve adding random noise. We demonstrate the limitations of existing frameworks and propose potential defense strategies analyzing the implications of using differential privacy. The novel insights presented in this study will aid in the improvement of FL frameworks. AVAILABILITY AND IMPLEMENTATION: The code examples are provided at GitHub (https://github.com/manuhuth/Data-Leakage-From-Covariances.git). The CNSIM1 dataset, which we used in the manuscript, is available within the DSData R package (https://github.com/datashield/DSData/tree/main/data). Oxford University Press 2023-08-30 /pmc/articles/PMC10516515/ /pubmed/37647639 http://dx.doi.org/10.1093/bioinformatics/btad531 Text en © The Author(s) 2023. Published by Oxford University Press. https://creativecommons.org/licenses/by/4.0/This is an Open Access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted reuse, distribution, and reproduction in any medium, provided the original work is properly cited. |
spellingShingle | Original Paper Huth, Manuel Arruda, Jonas Gusinow, Roy Contento, Lorenzo Tacconelli, Evelina Hasenauer, Jan Accessibility of covariance information creates vulnerability in Federated Learning frameworks |
title | Accessibility of covariance information creates vulnerability in Federated Learning frameworks |
title_full | Accessibility of covariance information creates vulnerability in Federated Learning frameworks |
title_fullStr | Accessibility of covariance information creates vulnerability in Federated Learning frameworks |
title_full_unstemmed | Accessibility of covariance information creates vulnerability in Federated Learning frameworks |
title_short | Accessibility of covariance information creates vulnerability in Federated Learning frameworks |
title_sort | accessibility of covariance information creates vulnerability in federated learning frameworks |
topic | Original Paper |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10516515/ https://www.ncbi.nlm.nih.gov/pubmed/37647639 http://dx.doi.org/10.1093/bioinformatics/btad531 |
work_keys_str_mv | AT huthmanuel accessibilityofcovarianceinformationcreatesvulnerabilityinfederatedlearningframeworks AT arrudajonas accessibilityofcovarianceinformationcreatesvulnerabilityinfederatedlearningframeworks AT gusinowroy accessibilityofcovarianceinformationcreatesvulnerabilityinfederatedlearningframeworks AT contentolorenzo accessibilityofcovarianceinformationcreatesvulnerabilityinfederatedlearningframeworks AT tacconellievelina accessibilityofcovarianceinformationcreatesvulnerabilityinfederatedlearningframeworks AT hasenauerjan accessibilityofcovarianceinformationcreatesvulnerabilityinfederatedlearningframeworks |