Cargando…

Detecting model misconducts in decentralized healthcare federated learning

BACKGROUND: To accelerate healthcare/genomic medicine research and facilitate quality improvement, researchers have started cross-institutional collaborations to use artificial intelligence on clinical/genomic data. However, there are real-world risks of incorrect models being submitted to the learn...

Descripción completa

Detalles Bibliográficos
Autores principales: Kuo, Tsung-Ting, Pham, Anh
Formato: Online Artículo Texto
Lenguaje:English
Publicado: 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10017272/
https://www.ncbi.nlm.nih.gov/pubmed/34923447
http://dx.doi.org/10.1016/j.ijmedinf.2021.104658
_version_ 1784907544440864768
author Kuo, Tsung-Ting
Pham, Anh
author_facet Kuo, Tsung-Ting
Pham, Anh
author_sort Kuo, Tsung-Ting
collection PubMed
description BACKGROUND: To accelerate healthcare/genomic medicine research and facilitate quality improvement, researchers have started cross-institutional collaborations to use artificial intelligence on clinical/genomic data. However, there are real-world risks of incorrect models being submitted to the learning process, due to either unforeseen accidents or malicious intent. This may reduce the incentives for institutions to participate in the federated modeling consortium. Existing methods to deal with this “model misconduct” issue mainly focus on modifying the learning methods, and therefore are more specifically tied with the algorithm. BASIC PROCEDURES: In this paper, we aim at solving the problem in an algorithm-agnostic way by (1) designing a simulator to generate various types of model misconduct, (2) developing a framework to detect the model misconducts, and (3) providing a generalizable approach to identify model misconducts for federated learning. We considered the following three categories: Plagiarism, Fabrication, and Falsification, and then developed a detection framework with three components: Auditing, Coefficient, and Performance detectors, with greedy parameter tuning. MAIN FINDINGS: We generated 10 types of misconducts from models learned on three datasets to evaluate our detection method. Our experiments showed high recall with low added computational cost. Our proposed detection method can best identify the misconduct on specific sites from any learning iteration, whereas it is more challenging to precisely detect misconducts for a specific site and at a specific iteration. PRINCIPAL CONCLUSIONS: We anticipate our study can support the enhancement of the integrity and reliability of federated machine learning on genomic/healthcare data.
format Online
Article
Text
id pubmed-10017272
institution National Center for Biotechnology Information
language English
publishDate 2021
record_format MEDLINE/PubMed
spelling pubmed-100172722023-06-09 Detecting model misconducts in decentralized healthcare federated learning Kuo, Tsung-Ting Pham, Anh Int J Med Inform Article BACKGROUND: To accelerate healthcare/genomic medicine research and facilitate quality improvement, researchers have started cross-institutional collaborations to use artificial intelligence on clinical/genomic data. However, there are real-world risks of incorrect models being submitted to the learning process, due to either unforeseen accidents or malicious intent. This may reduce the incentives for institutions to participate in the federated modeling consortium. Existing methods to deal with this “model misconduct” issue mainly focus on modifying the learning methods, and therefore are more specifically tied with the algorithm. BASIC PROCEDURES: In this paper, we aim at solving the problem in an algorithm-agnostic way by (1) designing a simulator to generate various types of model misconduct, (2) developing a framework to detect the model misconducts, and (3) providing a generalizable approach to identify model misconducts for federated learning. We considered the following three categories: Plagiarism, Fabrication, and Falsification, and then developed a detection framework with three components: Auditing, Coefficient, and Performance detectors, with greedy parameter tuning. MAIN FINDINGS: We generated 10 types of misconducts from models learned on three datasets to evaluate our detection method. Our experiments showed high recall with low added computational cost. Our proposed detection method can best identify the misconduct on specific sites from any learning iteration, whereas it is more challenging to precisely detect misconducts for a specific site and at a specific iteration. PRINCIPAL CONCLUSIONS: We anticipate our study can support the enhancement of the integrity and reliability of federated machine learning on genomic/healthcare data. 2021-12-09 /pmc/articles/PMC10017272/ /pubmed/34923447 http://dx.doi.org/10.1016/j.ijmedinf.2021.104658 Text en https://creativecommons.org/licenses/by/4.0/This is an open access article under the CC BY license (http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) ).
spellingShingle Article
Kuo, Tsung-Ting
Pham, Anh
Detecting model misconducts in decentralized healthcare federated learning
title Detecting model misconducts in decentralized healthcare federated learning
title_full Detecting model misconducts in decentralized healthcare federated learning
title_fullStr Detecting model misconducts in decentralized healthcare federated learning
title_full_unstemmed Detecting model misconducts in decentralized healthcare federated learning
title_short Detecting model misconducts in decentralized healthcare federated learning
title_sort detecting model misconducts in decentralized healthcare federated learning
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10017272/
https://www.ncbi.nlm.nih.gov/pubmed/34923447
http://dx.doi.org/10.1016/j.ijmedinf.2021.104658
work_keys_str_mv AT kuotsungting detectingmodelmisconductsindecentralizedhealthcarefederatedlearning
AT phamanh detectingmodelmisconductsindecentralizedhealthcarefederatedlearning