Cargando…
Differentially private knowledge transfer for federated learning
Extracting useful knowledge from big data is important for machine learning. When data is privacy-sensitive and cannot be directly collected, federated learning is a promising option that extracts knowledge from decentralized data by learning and exchanging model parameters, rather than raw data. Ho...
Autores principales: | , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Nature Publishing Group UK
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10290720/ https://www.ncbi.nlm.nih.gov/pubmed/37355643 http://dx.doi.org/10.1038/s41467-023-38794-x |
_version_ | 1785062552001052672 |
---|---|
author | Qi, Tao Wu, Fangzhao Wu, Chuhan He, Liang Huang, Yongfeng Xie, Xing |
author_facet | Qi, Tao Wu, Fangzhao Wu, Chuhan He, Liang Huang, Yongfeng Xie, Xing |
author_sort | Qi, Tao |
collection | PubMed |
description | Extracting useful knowledge from big data is important for machine learning. When data is privacy-sensitive and cannot be directly collected, federated learning is a promising option that extracts knowledge from decentralized data by learning and exchanging model parameters, rather than raw data. However, model parameters may encode not only non-private knowledge but also private information of local data, thereby transferring knowledge via model parameters is not privacy-secure. Here, we present a knowledge transfer method named PrivateKT, which uses actively selected small public data to transfer high-quality knowledge in federated learning with privacy guarantees. We verify PrivateKT on three different datasets, and results show that PrivateKT can maximally reduce 84% of the performance gap between centralized learning and existing federated learning methods under strict differential privacy restrictions. PrivateKT provides a potential direction to effective and privacy-preserving knowledge transfer in machine intelligent systems. |
format | Online Article Text |
id | pubmed-10290720 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Nature Publishing Group UK |
record_format | MEDLINE/PubMed |
spelling | pubmed-102907202023-06-26 Differentially private knowledge transfer for federated learning Qi, Tao Wu, Fangzhao Wu, Chuhan He, Liang Huang, Yongfeng Xie, Xing Nat Commun Article Extracting useful knowledge from big data is important for machine learning. When data is privacy-sensitive and cannot be directly collected, federated learning is a promising option that extracts knowledge from decentralized data by learning and exchanging model parameters, rather than raw data. However, model parameters may encode not only non-private knowledge but also private information of local data, thereby transferring knowledge via model parameters is not privacy-secure. Here, we present a knowledge transfer method named PrivateKT, which uses actively selected small public data to transfer high-quality knowledge in federated learning with privacy guarantees. We verify PrivateKT on three different datasets, and results show that PrivateKT can maximally reduce 84% of the performance gap between centralized learning and existing federated learning methods under strict differential privacy restrictions. PrivateKT provides a potential direction to effective and privacy-preserving knowledge transfer in machine intelligent systems. Nature Publishing Group UK 2023-06-24 /pmc/articles/PMC10290720/ /pubmed/37355643 http://dx.doi.org/10.1038/s41467-023-38794-x Text en © The Author(s) 2023 https://creativecommons.org/licenses/by/4.0/Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) . |
spellingShingle | Article Qi, Tao Wu, Fangzhao Wu, Chuhan He, Liang Huang, Yongfeng Xie, Xing Differentially private knowledge transfer for federated learning |
title | Differentially private knowledge transfer for federated learning |
title_full | Differentially private knowledge transfer for federated learning |
title_fullStr | Differentially private knowledge transfer for federated learning |
title_full_unstemmed | Differentially private knowledge transfer for federated learning |
title_short | Differentially private knowledge transfer for federated learning |
title_sort | differentially private knowledge transfer for federated learning |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10290720/ https://www.ncbi.nlm.nih.gov/pubmed/37355643 http://dx.doi.org/10.1038/s41467-023-38794-x |
work_keys_str_mv | AT qitao differentiallyprivateknowledgetransferforfederatedlearning AT wufangzhao differentiallyprivateknowledgetransferforfederatedlearning AT wuchuhan differentiallyprivateknowledgetransferforfederatedlearning AT heliang differentiallyprivateknowledgetransferforfederatedlearning AT huangyongfeng differentiallyprivateknowledgetransferforfederatedlearning AT xiexing differentiallyprivateknowledgetransferforfederatedlearning |