Cargando…

A study of generative large language model for medical research and healthcare

There are enormous enthusiasm and concerns in applying large language models (LLMs) to healthcare. Yet current assumptions are based on general-purpose LLMs such as ChatGPT, which are not developed for medical use. This study develops a generative clinical LLM, GatorTronGPT, using 277 billion words...

Descripción completa

Detalles Bibliográficos
Autores principales: Peng, Cheng, Yang, Xi, Chen, Aokun, Smith, Kaleb E., PourNejatian, Nima, Costa, Anthony B., Martin, Cheryl, Flores, Mona G., Zhang, Ying, Magoc, Tanja, Lipori, Gloria, Mitchell, Duane A., Ospina, Naykky S., Ahmed, Mustafa M., Hogan, William R., Shenkman, Elizabeth A., Guo, Yi, Bian, Jiang, Wu, Yonghui
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Nature Publishing Group UK 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10654385/
https://www.ncbi.nlm.nih.gov/pubmed/37973919
http://dx.doi.org/10.1038/s41746-023-00958-w
_version_ 1785147836978954240
author Peng, Cheng
Yang, Xi
Chen, Aokun
Smith, Kaleb E.
PourNejatian, Nima
Costa, Anthony B.
Martin, Cheryl
Flores, Mona G.
Zhang, Ying
Magoc, Tanja
Lipori, Gloria
Mitchell, Duane A.
Ospina, Naykky S.
Ahmed, Mustafa M.
Hogan, William R.
Shenkman, Elizabeth A.
Guo, Yi
Bian, Jiang
Wu, Yonghui
author_facet Peng, Cheng
Yang, Xi
Chen, Aokun
Smith, Kaleb E.
PourNejatian, Nima
Costa, Anthony B.
Martin, Cheryl
Flores, Mona G.
Zhang, Ying
Magoc, Tanja
Lipori, Gloria
Mitchell, Duane A.
Ospina, Naykky S.
Ahmed, Mustafa M.
Hogan, William R.
Shenkman, Elizabeth A.
Guo, Yi
Bian, Jiang
Wu, Yonghui
author_sort Peng, Cheng
collection PubMed
description There are enormous enthusiasm and concerns in applying large language models (LLMs) to healthcare. Yet current assumptions are based on general-purpose LLMs such as ChatGPT, which are not developed for medical use. This study develops a generative clinical LLM, GatorTronGPT, using 277 billion words of text including (1) 82 billion words of clinical text from 126 clinical departments and approximately 2 million patients at the University of Florida Health and (2) 195 billion words of diverse general English text. We train GatorTronGPT using a GPT-3 architecture with up to 20 billion parameters and evaluate its utility for biomedical natural language processing (NLP) and healthcare text generation. GatorTronGPT improves biomedical natural language processing. We apply GatorTronGPT to generate 20 billion words of synthetic text. Synthetic NLP models trained using synthetic text generated by GatorTronGPT outperform models trained using real-world clinical text. Physicians’ Turing test using 1 (worst) to 9 (best) scale shows that there are no significant differences in linguistic readability (p = 0.22; 6.57 of GatorTronGPT compared with 6.93 of human) and clinical relevance (p = 0.91; 7.0 of GatorTronGPT compared with 6.97 of human) and that physicians cannot differentiate them (p < 0.001). This study provides insights into the opportunities and challenges of LLMs for medical research and healthcare.
format Online
Article
Text
id pubmed-10654385
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Nature Publishing Group UK
record_format MEDLINE/PubMed
spelling pubmed-106543852023-11-16 A study of generative large language model for medical research and healthcare Peng, Cheng Yang, Xi Chen, Aokun Smith, Kaleb E. PourNejatian, Nima Costa, Anthony B. Martin, Cheryl Flores, Mona G. Zhang, Ying Magoc, Tanja Lipori, Gloria Mitchell, Duane A. Ospina, Naykky S. Ahmed, Mustafa M. Hogan, William R. Shenkman, Elizabeth A. Guo, Yi Bian, Jiang Wu, Yonghui NPJ Digit Med Article There are enormous enthusiasm and concerns in applying large language models (LLMs) to healthcare. Yet current assumptions are based on general-purpose LLMs such as ChatGPT, which are not developed for medical use. This study develops a generative clinical LLM, GatorTronGPT, using 277 billion words of text including (1) 82 billion words of clinical text from 126 clinical departments and approximately 2 million patients at the University of Florida Health and (2) 195 billion words of diverse general English text. We train GatorTronGPT using a GPT-3 architecture with up to 20 billion parameters and evaluate its utility for biomedical natural language processing (NLP) and healthcare text generation. GatorTronGPT improves biomedical natural language processing. We apply GatorTronGPT to generate 20 billion words of synthetic text. Synthetic NLP models trained using synthetic text generated by GatorTronGPT outperform models trained using real-world clinical text. Physicians’ Turing test using 1 (worst) to 9 (best) scale shows that there are no significant differences in linguistic readability (p = 0.22; 6.57 of GatorTronGPT compared with 6.93 of human) and clinical relevance (p = 0.91; 7.0 of GatorTronGPT compared with 6.97 of human) and that physicians cannot differentiate them (p < 0.001). This study provides insights into the opportunities and challenges of LLMs for medical research and healthcare. Nature Publishing Group UK 2023-11-16 /pmc/articles/PMC10654385/ /pubmed/37973919 http://dx.doi.org/10.1038/s41746-023-00958-w Text en © The Author(s) 2023 https://creativecommons.org/licenses/by/4.0/Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) .
spellingShingle Article
Peng, Cheng
Yang, Xi
Chen, Aokun
Smith, Kaleb E.
PourNejatian, Nima
Costa, Anthony B.
Martin, Cheryl
Flores, Mona G.
Zhang, Ying
Magoc, Tanja
Lipori, Gloria
Mitchell, Duane A.
Ospina, Naykky S.
Ahmed, Mustafa M.
Hogan, William R.
Shenkman, Elizabeth A.
Guo, Yi
Bian, Jiang
Wu, Yonghui
A study of generative large language model for medical research and healthcare
title A study of generative large language model for medical research and healthcare
title_full A study of generative large language model for medical research and healthcare
title_fullStr A study of generative large language model for medical research and healthcare
title_full_unstemmed A study of generative large language model for medical research and healthcare
title_short A study of generative large language model for medical research and healthcare
title_sort study of generative large language model for medical research and healthcare
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10654385/
https://www.ncbi.nlm.nih.gov/pubmed/37973919
http://dx.doi.org/10.1038/s41746-023-00958-w
work_keys_str_mv AT pengcheng astudyofgenerativelargelanguagemodelformedicalresearchandhealthcare
AT yangxi astudyofgenerativelargelanguagemodelformedicalresearchandhealthcare
AT chenaokun astudyofgenerativelargelanguagemodelformedicalresearchandhealthcare
AT smithkalebe astudyofgenerativelargelanguagemodelformedicalresearchandhealthcare
AT pournejatiannima astudyofgenerativelargelanguagemodelformedicalresearchandhealthcare
AT costaanthonyb astudyofgenerativelargelanguagemodelformedicalresearchandhealthcare
AT martincheryl astudyofgenerativelargelanguagemodelformedicalresearchandhealthcare
AT floresmonag astudyofgenerativelargelanguagemodelformedicalresearchandhealthcare
AT zhangying astudyofgenerativelargelanguagemodelformedicalresearchandhealthcare
AT magoctanja astudyofgenerativelargelanguagemodelformedicalresearchandhealthcare
AT liporigloria astudyofgenerativelargelanguagemodelformedicalresearchandhealthcare
AT mitchellduanea astudyofgenerativelargelanguagemodelformedicalresearchandhealthcare
AT ospinanaykkys astudyofgenerativelargelanguagemodelformedicalresearchandhealthcare
AT ahmedmustafam astudyofgenerativelargelanguagemodelformedicalresearchandhealthcare
AT hoganwilliamr astudyofgenerativelargelanguagemodelformedicalresearchandhealthcare
AT shenkmanelizabetha astudyofgenerativelargelanguagemodelformedicalresearchandhealthcare
AT guoyi astudyofgenerativelargelanguagemodelformedicalresearchandhealthcare
AT bianjiang astudyofgenerativelargelanguagemodelformedicalresearchandhealthcare
AT wuyonghui astudyofgenerativelargelanguagemodelformedicalresearchandhealthcare
AT pengcheng studyofgenerativelargelanguagemodelformedicalresearchandhealthcare
AT yangxi studyofgenerativelargelanguagemodelformedicalresearchandhealthcare
AT chenaokun studyofgenerativelargelanguagemodelformedicalresearchandhealthcare
AT smithkalebe studyofgenerativelargelanguagemodelformedicalresearchandhealthcare
AT pournejatiannima studyofgenerativelargelanguagemodelformedicalresearchandhealthcare
AT costaanthonyb studyofgenerativelargelanguagemodelformedicalresearchandhealthcare
AT martincheryl studyofgenerativelargelanguagemodelformedicalresearchandhealthcare
AT floresmonag studyofgenerativelargelanguagemodelformedicalresearchandhealthcare
AT zhangying studyofgenerativelargelanguagemodelformedicalresearchandhealthcare
AT magoctanja studyofgenerativelargelanguagemodelformedicalresearchandhealthcare
AT liporigloria studyofgenerativelargelanguagemodelformedicalresearchandhealthcare
AT mitchellduanea studyofgenerativelargelanguagemodelformedicalresearchandhealthcare
AT ospinanaykkys studyofgenerativelargelanguagemodelformedicalresearchandhealthcare
AT ahmedmustafam studyofgenerativelargelanguagemodelformedicalresearchandhealthcare
AT hoganwilliamr studyofgenerativelargelanguagemodelformedicalresearchandhealthcare
AT shenkmanelizabetha studyofgenerativelargelanguagemodelformedicalresearchandhealthcare
AT guoyi studyofgenerativelargelanguagemodelformedicalresearchandhealthcare
AT bianjiang studyofgenerativelargelanguagemodelformedicalresearchandhealthcare
AT wuyonghui studyofgenerativelargelanguagemodelformedicalresearchandhealthcare