Cargando…
A study of generative large language model for medical research and healthcare
There are enormous enthusiasm and concerns in applying large language models (LLMs) to healthcare. Yet current assumptions are based on general-purpose LLMs such as ChatGPT, which are not developed for medical use. This study develops a generative clinical LLM, GatorTronGPT, using 277 billion words...
Autores principales: | , , , , , , , , , , , , , , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Nature Publishing Group UK
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10654385/ https://www.ncbi.nlm.nih.gov/pubmed/37973919 http://dx.doi.org/10.1038/s41746-023-00958-w |
_version_ | 1785147836978954240 |
---|---|
author | Peng, Cheng Yang, Xi Chen, Aokun Smith, Kaleb E. PourNejatian, Nima Costa, Anthony B. Martin, Cheryl Flores, Mona G. Zhang, Ying Magoc, Tanja Lipori, Gloria Mitchell, Duane A. Ospina, Naykky S. Ahmed, Mustafa M. Hogan, William R. Shenkman, Elizabeth A. Guo, Yi Bian, Jiang Wu, Yonghui |
author_facet | Peng, Cheng Yang, Xi Chen, Aokun Smith, Kaleb E. PourNejatian, Nima Costa, Anthony B. Martin, Cheryl Flores, Mona G. Zhang, Ying Magoc, Tanja Lipori, Gloria Mitchell, Duane A. Ospina, Naykky S. Ahmed, Mustafa M. Hogan, William R. Shenkman, Elizabeth A. Guo, Yi Bian, Jiang Wu, Yonghui |
author_sort | Peng, Cheng |
collection | PubMed |
description | There are enormous enthusiasm and concerns in applying large language models (LLMs) to healthcare. Yet current assumptions are based on general-purpose LLMs such as ChatGPT, which are not developed for medical use. This study develops a generative clinical LLM, GatorTronGPT, using 277 billion words of text including (1) 82 billion words of clinical text from 126 clinical departments and approximately 2 million patients at the University of Florida Health and (2) 195 billion words of diverse general English text. We train GatorTronGPT using a GPT-3 architecture with up to 20 billion parameters and evaluate its utility for biomedical natural language processing (NLP) and healthcare text generation. GatorTronGPT improves biomedical natural language processing. We apply GatorTronGPT to generate 20 billion words of synthetic text. Synthetic NLP models trained using synthetic text generated by GatorTronGPT outperform models trained using real-world clinical text. Physicians’ Turing test using 1 (worst) to 9 (best) scale shows that there are no significant differences in linguistic readability (p = 0.22; 6.57 of GatorTronGPT compared with 6.93 of human) and clinical relevance (p = 0.91; 7.0 of GatorTronGPT compared with 6.97 of human) and that physicians cannot differentiate them (p < 0.001). This study provides insights into the opportunities and challenges of LLMs for medical research and healthcare. |
format | Online Article Text |
id | pubmed-10654385 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Nature Publishing Group UK |
record_format | MEDLINE/PubMed |
spelling | pubmed-106543852023-11-16 A study of generative large language model for medical research and healthcare Peng, Cheng Yang, Xi Chen, Aokun Smith, Kaleb E. PourNejatian, Nima Costa, Anthony B. Martin, Cheryl Flores, Mona G. Zhang, Ying Magoc, Tanja Lipori, Gloria Mitchell, Duane A. Ospina, Naykky S. Ahmed, Mustafa M. Hogan, William R. Shenkman, Elizabeth A. Guo, Yi Bian, Jiang Wu, Yonghui NPJ Digit Med Article There are enormous enthusiasm and concerns in applying large language models (LLMs) to healthcare. Yet current assumptions are based on general-purpose LLMs such as ChatGPT, which are not developed for medical use. This study develops a generative clinical LLM, GatorTronGPT, using 277 billion words of text including (1) 82 billion words of clinical text from 126 clinical departments and approximately 2 million patients at the University of Florida Health and (2) 195 billion words of diverse general English text. We train GatorTronGPT using a GPT-3 architecture with up to 20 billion parameters and evaluate its utility for biomedical natural language processing (NLP) and healthcare text generation. GatorTronGPT improves biomedical natural language processing. We apply GatorTronGPT to generate 20 billion words of synthetic text. Synthetic NLP models trained using synthetic text generated by GatorTronGPT outperform models trained using real-world clinical text. Physicians’ Turing test using 1 (worst) to 9 (best) scale shows that there are no significant differences in linguistic readability (p = 0.22; 6.57 of GatorTronGPT compared with 6.93 of human) and clinical relevance (p = 0.91; 7.0 of GatorTronGPT compared with 6.97 of human) and that physicians cannot differentiate them (p < 0.001). This study provides insights into the opportunities and challenges of LLMs for medical research and healthcare. Nature Publishing Group UK 2023-11-16 /pmc/articles/PMC10654385/ /pubmed/37973919 http://dx.doi.org/10.1038/s41746-023-00958-w Text en © The Author(s) 2023 https://creativecommons.org/licenses/by/4.0/Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) . |
spellingShingle | Article Peng, Cheng Yang, Xi Chen, Aokun Smith, Kaleb E. PourNejatian, Nima Costa, Anthony B. Martin, Cheryl Flores, Mona G. Zhang, Ying Magoc, Tanja Lipori, Gloria Mitchell, Duane A. Ospina, Naykky S. Ahmed, Mustafa M. Hogan, William R. Shenkman, Elizabeth A. Guo, Yi Bian, Jiang Wu, Yonghui A study of generative large language model for medical research and healthcare |
title | A study of generative large language model for medical research and healthcare |
title_full | A study of generative large language model for medical research and healthcare |
title_fullStr | A study of generative large language model for medical research and healthcare |
title_full_unstemmed | A study of generative large language model for medical research and healthcare |
title_short | A study of generative large language model for medical research and healthcare |
title_sort | study of generative large language model for medical research and healthcare |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10654385/ https://www.ncbi.nlm.nih.gov/pubmed/37973919 http://dx.doi.org/10.1038/s41746-023-00958-w |
work_keys_str_mv | AT pengcheng astudyofgenerativelargelanguagemodelformedicalresearchandhealthcare AT yangxi astudyofgenerativelargelanguagemodelformedicalresearchandhealthcare AT chenaokun astudyofgenerativelargelanguagemodelformedicalresearchandhealthcare AT smithkalebe astudyofgenerativelargelanguagemodelformedicalresearchandhealthcare AT pournejatiannima astudyofgenerativelargelanguagemodelformedicalresearchandhealthcare AT costaanthonyb astudyofgenerativelargelanguagemodelformedicalresearchandhealthcare AT martincheryl astudyofgenerativelargelanguagemodelformedicalresearchandhealthcare AT floresmonag astudyofgenerativelargelanguagemodelformedicalresearchandhealthcare AT zhangying astudyofgenerativelargelanguagemodelformedicalresearchandhealthcare AT magoctanja astudyofgenerativelargelanguagemodelformedicalresearchandhealthcare AT liporigloria astudyofgenerativelargelanguagemodelformedicalresearchandhealthcare AT mitchellduanea astudyofgenerativelargelanguagemodelformedicalresearchandhealthcare AT ospinanaykkys astudyofgenerativelargelanguagemodelformedicalresearchandhealthcare AT ahmedmustafam astudyofgenerativelargelanguagemodelformedicalresearchandhealthcare AT hoganwilliamr astudyofgenerativelargelanguagemodelformedicalresearchandhealthcare AT shenkmanelizabetha astudyofgenerativelargelanguagemodelformedicalresearchandhealthcare AT guoyi astudyofgenerativelargelanguagemodelformedicalresearchandhealthcare AT bianjiang astudyofgenerativelargelanguagemodelformedicalresearchandhealthcare AT wuyonghui astudyofgenerativelargelanguagemodelformedicalresearchandhealthcare AT pengcheng studyofgenerativelargelanguagemodelformedicalresearchandhealthcare AT yangxi studyofgenerativelargelanguagemodelformedicalresearchandhealthcare AT chenaokun studyofgenerativelargelanguagemodelformedicalresearchandhealthcare AT smithkalebe studyofgenerativelargelanguagemodelformedicalresearchandhealthcare AT pournejatiannima studyofgenerativelargelanguagemodelformedicalresearchandhealthcare AT costaanthonyb studyofgenerativelargelanguagemodelformedicalresearchandhealthcare AT martincheryl studyofgenerativelargelanguagemodelformedicalresearchandhealthcare AT floresmonag studyofgenerativelargelanguagemodelformedicalresearchandhealthcare AT zhangying studyofgenerativelargelanguagemodelformedicalresearchandhealthcare AT magoctanja studyofgenerativelargelanguagemodelformedicalresearchandhealthcare AT liporigloria studyofgenerativelargelanguagemodelformedicalresearchandhealthcare AT mitchellduanea studyofgenerativelargelanguagemodelformedicalresearchandhealthcare AT ospinanaykkys studyofgenerativelargelanguagemodelformedicalresearchandhealthcare AT ahmedmustafam studyofgenerativelargelanguagemodelformedicalresearchandhealthcare AT hoganwilliamr studyofgenerativelargelanguagemodelformedicalresearchandhealthcare AT shenkmanelizabetha studyofgenerativelargelanguagemodelformedicalresearchandhealthcare AT guoyi studyofgenerativelargelanguagemodelformedicalresearchandhealthcare AT bianjiang studyofgenerativelargelanguagemodelformedicalresearchandhealthcare AT wuyonghui studyofgenerativelargelanguagemodelformedicalresearchandhealthcare |