Cargando…

The shaky foundations of large language models and foundation models for electronic health records

The success of foundation models such as ChatGPT and AlphaFold has spurred significant interest in building similar models for electronic medical records (EMRs) to improve patient care and hospital operations. However, recent hype has obscured critical gaps in our understanding of these models’ capa...

Descripción completa

Detalles Bibliográficos
Autores principales: Wornow, Michael, Xu, Yizhe, Thapa, Rahul, Patel, Birju, Steinberg, Ethan, Fleming, Scott, Pfeffer, Michael A., Fries, Jason, Shah, Nigam H.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Nature Publishing Group UK 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10387101/
https://www.ncbi.nlm.nih.gov/pubmed/37516790
http://dx.doi.org/10.1038/s41746-023-00879-8
_version_ 1785081814155526144
author Wornow, Michael
Xu, Yizhe
Thapa, Rahul
Patel, Birju
Steinberg, Ethan
Fleming, Scott
Pfeffer, Michael A.
Fries, Jason
Shah, Nigam H.
author_facet Wornow, Michael
Xu, Yizhe
Thapa, Rahul
Patel, Birju
Steinberg, Ethan
Fleming, Scott
Pfeffer, Michael A.
Fries, Jason
Shah, Nigam H.
author_sort Wornow, Michael
collection PubMed
description The success of foundation models such as ChatGPT and AlphaFold has spurred significant interest in building similar models for electronic medical records (EMRs) to improve patient care and hospital operations. However, recent hype has obscured critical gaps in our understanding of these models’ capabilities. In this narrative review, we examine 84 foundation models trained on non-imaging EMR data (i.e., clinical text and/or structured data) and create a taxonomy delineating their architectures, training data, and potential use cases. We find that most models are trained on small, narrowly-scoped clinical datasets (e.g., MIMIC-III) or broad, public biomedical corpora (e.g., PubMed) and are evaluated on tasks that do not provide meaningful insights on their usefulness to health systems. Considering these findings, we propose an improved evaluation framework for measuring the benefits of clinical foundation models that is more closely grounded to metrics that matter in healthcare.
format Online
Article
Text
id pubmed-10387101
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Nature Publishing Group UK
record_format MEDLINE/PubMed
spelling pubmed-103871012023-07-31 The shaky foundations of large language models and foundation models for electronic health records Wornow, Michael Xu, Yizhe Thapa, Rahul Patel, Birju Steinberg, Ethan Fleming, Scott Pfeffer, Michael A. Fries, Jason Shah, Nigam H. NPJ Digit Med Review Article The success of foundation models such as ChatGPT and AlphaFold has spurred significant interest in building similar models for electronic medical records (EMRs) to improve patient care and hospital operations. However, recent hype has obscured critical gaps in our understanding of these models’ capabilities. In this narrative review, we examine 84 foundation models trained on non-imaging EMR data (i.e., clinical text and/or structured data) and create a taxonomy delineating their architectures, training data, and potential use cases. We find that most models are trained on small, narrowly-scoped clinical datasets (e.g., MIMIC-III) or broad, public biomedical corpora (e.g., PubMed) and are evaluated on tasks that do not provide meaningful insights on their usefulness to health systems. Considering these findings, we propose an improved evaluation framework for measuring the benefits of clinical foundation models that is more closely grounded to metrics that matter in healthcare. Nature Publishing Group UK 2023-07-29 /pmc/articles/PMC10387101/ /pubmed/37516790 http://dx.doi.org/10.1038/s41746-023-00879-8 Text en © The Author(s) 2023 https://creativecommons.org/licenses/by/4.0/Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) .
spellingShingle Review Article
Wornow, Michael
Xu, Yizhe
Thapa, Rahul
Patel, Birju
Steinberg, Ethan
Fleming, Scott
Pfeffer, Michael A.
Fries, Jason
Shah, Nigam H.
The shaky foundations of large language models and foundation models for electronic health records
title The shaky foundations of large language models and foundation models for electronic health records
title_full The shaky foundations of large language models and foundation models for electronic health records
title_fullStr The shaky foundations of large language models and foundation models for electronic health records
title_full_unstemmed The shaky foundations of large language models and foundation models for electronic health records
title_short The shaky foundations of large language models and foundation models for electronic health records
title_sort shaky foundations of large language models and foundation models for electronic health records
topic Review Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10387101/
https://www.ncbi.nlm.nih.gov/pubmed/37516790
http://dx.doi.org/10.1038/s41746-023-00879-8
work_keys_str_mv AT wornowmichael theshakyfoundationsoflargelanguagemodelsandfoundationmodelsforelectronichealthrecords
AT xuyizhe theshakyfoundationsoflargelanguagemodelsandfoundationmodelsforelectronichealthrecords
AT thaparahul theshakyfoundationsoflargelanguagemodelsandfoundationmodelsforelectronichealthrecords
AT patelbirju theshakyfoundationsoflargelanguagemodelsandfoundationmodelsforelectronichealthrecords
AT steinbergethan theshakyfoundationsoflargelanguagemodelsandfoundationmodelsforelectronichealthrecords
AT flemingscott theshakyfoundationsoflargelanguagemodelsandfoundationmodelsforelectronichealthrecords
AT pfeffermichaela theshakyfoundationsoflargelanguagemodelsandfoundationmodelsforelectronichealthrecords
AT friesjason theshakyfoundationsoflargelanguagemodelsandfoundationmodelsforelectronichealthrecords
AT shahnigamh theshakyfoundationsoflargelanguagemodelsandfoundationmodelsforelectronichealthrecords
AT wornowmichael shakyfoundationsoflargelanguagemodelsandfoundationmodelsforelectronichealthrecords
AT xuyizhe shakyfoundationsoflargelanguagemodelsandfoundationmodelsforelectronichealthrecords
AT thaparahul shakyfoundationsoflargelanguagemodelsandfoundationmodelsforelectronichealthrecords
AT patelbirju shakyfoundationsoflargelanguagemodelsandfoundationmodelsforelectronichealthrecords
AT steinbergethan shakyfoundationsoflargelanguagemodelsandfoundationmodelsforelectronichealthrecords
AT flemingscott shakyfoundationsoflargelanguagemodelsandfoundationmodelsforelectronichealthrecords
AT pfeffermichaela shakyfoundationsoflargelanguagemodelsandfoundationmodelsforelectronichealthrecords
AT friesjason shakyfoundationsoflargelanguagemodelsandfoundationmodelsforelectronichealthrecords
AT shahnigamh shakyfoundationsoflargelanguagemodelsandfoundationmodelsforelectronichealthrecords