Cargando…
The shaky foundations of large language models and foundation models for electronic health records
The success of foundation models such as ChatGPT and AlphaFold has spurred significant interest in building similar models for electronic medical records (EMRs) to improve patient care and hospital operations. However, recent hype has obscured critical gaps in our understanding of these models’ capa...
Autores principales: | , , , , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Nature Publishing Group UK
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10387101/ https://www.ncbi.nlm.nih.gov/pubmed/37516790 http://dx.doi.org/10.1038/s41746-023-00879-8 |
_version_ | 1785081814155526144 |
---|---|
author | Wornow, Michael Xu, Yizhe Thapa, Rahul Patel, Birju Steinberg, Ethan Fleming, Scott Pfeffer, Michael A. Fries, Jason Shah, Nigam H. |
author_facet | Wornow, Michael Xu, Yizhe Thapa, Rahul Patel, Birju Steinberg, Ethan Fleming, Scott Pfeffer, Michael A. Fries, Jason Shah, Nigam H. |
author_sort | Wornow, Michael |
collection | PubMed |
description | The success of foundation models such as ChatGPT and AlphaFold has spurred significant interest in building similar models for electronic medical records (EMRs) to improve patient care and hospital operations. However, recent hype has obscured critical gaps in our understanding of these models’ capabilities. In this narrative review, we examine 84 foundation models trained on non-imaging EMR data (i.e., clinical text and/or structured data) and create a taxonomy delineating their architectures, training data, and potential use cases. We find that most models are trained on small, narrowly-scoped clinical datasets (e.g., MIMIC-III) or broad, public biomedical corpora (e.g., PubMed) and are evaluated on tasks that do not provide meaningful insights on their usefulness to health systems. Considering these findings, we propose an improved evaluation framework for measuring the benefits of clinical foundation models that is more closely grounded to metrics that matter in healthcare. |
format | Online Article Text |
id | pubmed-10387101 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Nature Publishing Group UK |
record_format | MEDLINE/PubMed |
spelling | pubmed-103871012023-07-31 The shaky foundations of large language models and foundation models for electronic health records Wornow, Michael Xu, Yizhe Thapa, Rahul Patel, Birju Steinberg, Ethan Fleming, Scott Pfeffer, Michael A. Fries, Jason Shah, Nigam H. NPJ Digit Med Review Article The success of foundation models such as ChatGPT and AlphaFold has spurred significant interest in building similar models for electronic medical records (EMRs) to improve patient care and hospital operations. However, recent hype has obscured critical gaps in our understanding of these models’ capabilities. In this narrative review, we examine 84 foundation models trained on non-imaging EMR data (i.e., clinical text and/or structured data) and create a taxonomy delineating their architectures, training data, and potential use cases. We find that most models are trained on small, narrowly-scoped clinical datasets (e.g., MIMIC-III) or broad, public biomedical corpora (e.g., PubMed) and are evaluated on tasks that do not provide meaningful insights on their usefulness to health systems. Considering these findings, we propose an improved evaluation framework for measuring the benefits of clinical foundation models that is more closely grounded to metrics that matter in healthcare. Nature Publishing Group UK 2023-07-29 /pmc/articles/PMC10387101/ /pubmed/37516790 http://dx.doi.org/10.1038/s41746-023-00879-8 Text en © The Author(s) 2023 https://creativecommons.org/licenses/by/4.0/Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) . |
spellingShingle | Review Article Wornow, Michael Xu, Yizhe Thapa, Rahul Patel, Birju Steinberg, Ethan Fleming, Scott Pfeffer, Michael A. Fries, Jason Shah, Nigam H. The shaky foundations of large language models and foundation models for electronic health records |
title | The shaky foundations of large language models and foundation models for electronic health records |
title_full | The shaky foundations of large language models and foundation models for electronic health records |
title_fullStr | The shaky foundations of large language models and foundation models for electronic health records |
title_full_unstemmed | The shaky foundations of large language models and foundation models for electronic health records |
title_short | The shaky foundations of large language models and foundation models for electronic health records |
title_sort | shaky foundations of large language models and foundation models for electronic health records |
topic | Review Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10387101/ https://www.ncbi.nlm.nih.gov/pubmed/37516790 http://dx.doi.org/10.1038/s41746-023-00879-8 |
work_keys_str_mv | AT wornowmichael theshakyfoundationsoflargelanguagemodelsandfoundationmodelsforelectronichealthrecords AT xuyizhe theshakyfoundationsoflargelanguagemodelsandfoundationmodelsforelectronichealthrecords AT thaparahul theshakyfoundationsoflargelanguagemodelsandfoundationmodelsforelectronichealthrecords AT patelbirju theshakyfoundationsoflargelanguagemodelsandfoundationmodelsforelectronichealthrecords AT steinbergethan theshakyfoundationsoflargelanguagemodelsandfoundationmodelsforelectronichealthrecords AT flemingscott theshakyfoundationsoflargelanguagemodelsandfoundationmodelsforelectronichealthrecords AT pfeffermichaela theshakyfoundationsoflargelanguagemodelsandfoundationmodelsforelectronichealthrecords AT friesjason theshakyfoundationsoflargelanguagemodelsandfoundationmodelsforelectronichealthrecords AT shahnigamh theshakyfoundationsoflargelanguagemodelsandfoundationmodelsforelectronichealthrecords AT wornowmichael shakyfoundationsoflargelanguagemodelsandfoundationmodelsforelectronichealthrecords AT xuyizhe shakyfoundationsoflargelanguagemodelsandfoundationmodelsforelectronichealthrecords AT thaparahul shakyfoundationsoflargelanguagemodelsandfoundationmodelsforelectronichealthrecords AT patelbirju shakyfoundationsoflargelanguagemodelsandfoundationmodelsforelectronichealthrecords AT steinbergethan shakyfoundationsoflargelanguagemodelsandfoundationmodelsforelectronichealthrecords AT flemingscott shakyfoundationsoflargelanguagemodelsandfoundationmodelsforelectronichealthrecords AT pfeffermichaela shakyfoundationsoflargelanguagemodelsandfoundationmodelsforelectronichealthrecords AT friesjason shakyfoundationsoflargelanguagemodelsandfoundationmodelsforelectronichealthrecords AT shahnigamh shakyfoundationsoflargelanguagemodelsandfoundationmodelsforelectronichealthrecords |