Cargando…
Federated learning improves site performance in multicenter deep learning without data sharing
OBJECTIVE: To demonstrate enabling multi-institutional training without centralizing or sharing the underlying physical data via federated learning (FL). MATERIALS AND METHODS: Deep learning models were trained at each participating institution using local clinical data, and an additional model was...
Autores principales: | , , , , , , , , , , , , , , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Oxford University Press
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8200268/ https://www.ncbi.nlm.nih.gov/pubmed/33537772 http://dx.doi.org/10.1093/jamia/ocaa341 |
_version_ | 1783707570213486592 |
---|---|
author | Sarma, Karthik V Harmon, Stephanie Sanford, Thomas Roth, Holger R Xu, Ziyue Tetreault, Jesse Xu, Daguang Flores, Mona G Raman, Alex G Kulkarni, Rushikesh Wood, Bradford J Choyke, Peter L Priester, Alan M Marks, Leonard S Raman, Steven S Enzmann, Dieter Turkbey, Baris Speier, William Arnold, Corey W |
author_facet | Sarma, Karthik V Harmon, Stephanie Sanford, Thomas Roth, Holger R Xu, Ziyue Tetreault, Jesse Xu, Daguang Flores, Mona G Raman, Alex G Kulkarni, Rushikesh Wood, Bradford J Choyke, Peter L Priester, Alan M Marks, Leonard S Raman, Steven S Enzmann, Dieter Turkbey, Baris Speier, William Arnold, Corey W |
author_sort | Sarma, Karthik V |
collection | PubMed |
description | OBJECTIVE: To demonstrate enabling multi-institutional training without centralizing or sharing the underlying physical data via federated learning (FL). MATERIALS AND METHODS: Deep learning models were trained at each participating institution using local clinical data, and an additional model was trained using FL across all of the institutions. RESULTS: We found that the FL model exhibited superior performance and generalizability to the models trained at single institutions, with an overall performance level that was significantly better than that of any of the institutional models alone when evaluated on held-out test sets from each institution and an outside challenge dataset. DISCUSSION: The power of FL was successfully demonstrated across 3 academic institutions while avoiding the privacy risk associated with the transfer and pooling of patient data. CONCLUSION: Federated learning is an effective methodology that merits further study to enable accelerated development of models across institutions, enabling greater generalizability in clinical use. |
format | Online Article Text |
id | pubmed-8200268 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | Oxford University Press |
record_format | MEDLINE/PubMed |
spelling | pubmed-82002682021-06-14 Federated learning improves site performance in multicenter deep learning without data sharing Sarma, Karthik V Harmon, Stephanie Sanford, Thomas Roth, Holger R Xu, Ziyue Tetreault, Jesse Xu, Daguang Flores, Mona G Raman, Alex G Kulkarni, Rushikesh Wood, Bradford J Choyke, Peter L Priester, Alan M Marks, Leonard S Raman, Steven S Enzmann, Dieter Turkbey, Baris Speier, William Arnold, Corey W J Am Med Inform Assoc Brief Communications OBJECTIVE: To demonstrate enabling multi-institutional training without centralizing or sharing the underlying physical data via federated learning (FL). MATERIALS AND METHODS: Deep learning models were trained at each participating institution using local clinical data, and an additional model was trained using FL across all of the institutions. RESULTS: We found that the FL model exhibited superior performance and generalizability to the models trained at single institutions, with an overall performance level that was significantly better than that of any of the institutional models alone when evaluated on held-out test sets from each institution and an outside challenge dataset. DISCUSSION: The power of FL was successfully demonstrated across 3 academic institutions while avoiding the privacy risk associated with the transfer and pooling of patient data. CONCLUSION: Federated learning is an effective methodology that merits further study to enable accelerated development of models across institutions, enabling greater generalizability in clinical use. Oxford University Press 2021-02-04 /pmc/articles/PMC8200268/ /pubmed/33537772 http://dx.doi.org/10.1093/jamia/ocaa341 Text en © The Author(s) 2021. Published by Oxford University Press on behalf of the American Medical Informatics Association. https://creativecommons.org/licenses/by/4.0/This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) ), which permits unrestricted reuse, distribution, and reproduction in any medium, provided the original work is properly cited. |
spellingShingle | Brief Communications Sarma, Karthik V Harmon, Stephanie Sanford, Thomas Roth, Holger R Xu, Ziyue Tetreault, Jesse Xu, Daguang Flores, Mona G Raman, Alex G Kulkarni, Rushikesh Wood, Bradford J Choyke, Peter L Priester, Alan M Marks, Leonard S Raman, Steven S Enzmann, Dieter Turkbey, Baris Speier, William Arnold, Corey W Federated learning improves site performance in multicenter deep learning without data sharing |
title | Federated learning improves site performance in multicenter deep learning without data sharing |
title_full | Federated learning improves site performance in multicenter deep learning without data sharing |
title_fullStr | Federated learning improves site performance in multicenter deep learning without data sharing |
title_full_unstemmed | Federated learning improves site performance in multicenter deep learning without data sharing |
title_short | Federated learning improves site performance in multicenter deep learning without data sharing |
title_sort | federated learning improves site performance in multicenter deep learning without data sharing |
topic | Brief Communications |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8200268/ https://www.ncbi.nlm.nih.gov/pubmed/33537772 http://dx.doi.org/10.1093/jamia/ocaa341 |
work_keys_str_mv | AT sarmakarthikv federatedlearningimprovessiteperformanceinmulticenterdeeplearningwithoutdatasharing AT harmonstephanie federatedlearningimprovessiteperformanceinmulticenterdeeplearningwithoutdatasharing AT sanfordthomas federatedlearningimprovessiteperformanceinmulticenterdeeplearningwithoutdatasharing AT rothholgerr federatedlearningimprovessiteperformanceinmulticenterdeeplearningwithoutdatasharing AT xuziyue federatedlearningimprovessiteperformanceinmulticenterdeeplearningwithoutdatasharing AT tetreaultjesse federatedlearningimprovessiteperformanceinmulticenterdeeplearningwithoutdatasharing AT xudaguang federatedlearningimprovessiteperformanceinmulticenterdeeplearningwithoutdatasharing AT floresmonag federatedlearningimprovessiteperformanceinmulticenterdeeplearningwithoutdatasharing AT ramanalexg federatedlearningimprovessiteperformanceinmulticenterdeeplearningwithoutdatasharing AT kulkarnirushikesh federatedlearningimprovessiteperformanceinmulticenterdeeplearningwithoutdatasharing AT woodbradfordj federatedlearningimprovessiteperformanceinmulticenterdeeplearningwithoutdatasharing AT choykepeterl federatedlearningimprovessiteperformanceinmulticenterdeeplearningwithoutdatasharing AT priesteralanm federatedlearningimprovessiteperformanceinmulticenterdeeplearningwithoutdatasharing AT marksleonards federatedlearningimprovessiteperformanceinmulticenterdeeplearningwithoutdatasharing AT ramanstevens federatedlearningimprovessiteperformanceinmulticenterdeeplearningwithoutdatasharing AT enzmanndieter federatedlearningimprovessiteperformanceinmulticenterdeeplearningwithoutdatasharing AT turkbeybaris federatedlearningimprovessiteperformanceinmulticenterdeeplearningwithoutdatasharing AT speierwilliam federatedlearningimprovessiteperformanceinmulticenterdeeplearningwithoutdatasharing AT arnoldcoreyw federatedlearningimprovessiteperformanceinmulticenterdeeplearningwithoutdatasharing |