Cargando…
Detecting model misconducts in decentralized healthcare federated learning
BACKGROUND: To accelerate healthcare/genomic medicine research and facilitate quality improvement, researchers have started cross-institutional collaborations to use artificial intelligence on clinical/genomic data. However, there are real-world risks of incorrect models being submitted to the learn...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10017272/ https://www.ncbi.nlm.nih.gov/pubmed/34923447 http://dx.doi.org/10.1016/j.ijmedinf.2021.104658 |
_version_ | 1784907544440864768 |
---|---|
author | Kuo, Tsung-Ting Pham, Anh |
author_facet | Kuo, Tsung-Ting Pham, Anh |
author_sort | Kuo, Tsung-Ting |
collection | PubMed |
description | BACKGROUND: To accelerate healthcare/genomic medicine research and facilitate quality improvement, researchers have started cross-institutional collaborations to use artificial intelligence on clinical/genomic data. However, there are real-world risks of incorrect models being submitted to the learning process, due to either unforeseen accidents or malicious intent. This may reduce the incentives for institutions to participate in the federated modeling consortium. Existing methods to deal with this “model misconduct” issue mainly focus on modifying the learning methods, and therefore are more specifically tied with the algorithm. BASIC PROCEDURES: In this paper, we aim at solving the problem in an algorithm-agnostic way by (1) designing a simulator to generate various types of model misconduct, (2) developing a framework to detect the model misconducts, and (3) providing a generalizable approach to identify model misconducts for federated learning. We considered the following three categories: Plagiarism, Fabrication, and Falsification, and then developed a detection framework with three components: Auditing, Coefficient, and Performance detectors, with greedy parameter tuning. MAIN FINDINGS: We generated 10 types of misconducts from models learned on three datasets to evaluate our detection method. Our experiments showed high recall with low added computational cost. Our proposed detection method can best identify the misconduct on specific sites from any learning iteration, whereas it is more challenging to precisely detect misconducts for a specific site and at a specific iteration. PRINCIPAL CONCLUSIONS: We anticipate our study can support the enhancement of the integrity and reliability of federated machine learning on genomic/healthcare data. |
format | Online Article Text |
id | pubmed-10017272 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
record_format | MEDLINE/PubMed |
spelling | pubmed-100172722023-06-09 Detecting model misconducts in decentralized healthcare federated learning Kuo, Tsung-Ting Pham, Anh Int J Med Inform Article BACKGROUND: To accelerate healthcare/genomic medicine research and facilitate quality improvement, researchers have started cross-institutional collaborations to use artificial intelligence on clinical/genomic data. However, there are real-world risks of incorrect models being submitted to the learning process, due to either unforeseen accidents or malicious intent. This may reduce the incentives for institutions to participate in the federated modeling consortium. Existing methods to deal with this “model misconduct” issue mainly focus on modifying the learning methods, and therefore are more specifically tied with the algorithm. BASIC PROCEDURES: In this paper, we aim at solving the problem in an algorithm-agnostic way by (1) designing a simulator to generate various types of model misconduct, (2) developing a framework to detect the model misconducts, and (3) providing a generalizable approach to identify model misconducts for federated learning. We considered the following three categories: Plagiarism, Fabrication, and Falsification, and then developed a detection framework with three components: Auditing, Coefficient, and Performance detectors, with greedy parameter tuning. MAIN FINDINGS: We generated 10 types of misconducts from models learned on three datasets to evaluate our detection method. Our experiments showed high recall with low added computational cost. Our proposed detection method can best identify the misconduct on specific sites from any learning iteration, whereas it is more challenging to precisely detect misconducts for a specific site and at a specific iteration. PRINCIPAL CONCLUSIONS: We anticipate our study can support the enhancement of the integrity and reliability of federated machine learning on genomic/healthcare data. 2021-12-09 /pmc/articles/PMC10017272/ /pubmed/34923447 http://dx.doi.org/10.1016/j.ijmedinf.2021.104658 Text en https://creativecommons.org/licenses/by/4.0/This is an open access article under the CC BY license (http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) ). |
spellingShingle | Article Kuo, Tsung-Ting Pham, Anh Detecting model misconducts in decentralized healthcare federated learning |
title | Detecting model misconducts in decentralized healthcare federated learning |
title_full | Detecting model misconducts in decentralized healthcare federated learning |
title_fullStr | Detecting model misconducts in decentralized healthcare federated learning |
title_full_unstemmed | Detecting model misconducts in decentralized healthcare federated learning |
title_short | Detecting model misconducts in decentralized healthcare federated learning |
title_sort | detecting model misconducts in decentralized healthcare federated learning |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10017272/ https://www.ncbi.nlm.nih.gov/pubmed/34923447 http://dx.doi.org/10.1016/j.ijmedinf.2021.104658 |
work_keys_str_mv | AT kuotsungting detectingmodelmisconductsindecentralizedhealthcarefederatedlearning AT phamanh detectingmodelmisconductsindecentralizedhealthcarefederatedlearning |