Cargando…
Machine learning for patient risk stratification: standing on, or looking over, the shoulders of clinicians?
Machine learning can help clinicians to make individualized patient predictions only if researchers demonstrate models that contribute novel insights, rather than learning the most likely next step in a set of actions a clinician will take. We trained deep learning models using only clinician-initia...
Autores principales: | , , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Nature Publishing Group UK
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8010071/ https://www.ncbi.nlm.nih.gov/pubmed/33785839 http://dx.doi.org/10.1038/s41746-021-00426-3 |
_version_ | 1783672986230849536 |
---|---|
author | Beaulieu-Jones, Brett K. Yuan, William Brat, Gabriel A. Beam, Andrew L. Weber, Griffin Ruffin, Marshall Kohane, Isaac S. |
author_facet | Beaulieu-Jones, Brett K. Yuan, William Brat, Gabriel A. Beam, Andrew L. Weber, Griffin Ruffin, Marshall Kohane, Isaac S. |
author_sort | Beaulieu-Jones, Brett K. |
collection | PubMed |
description | Machine learning can help clinicians to make individualized patient predictions only if researchers demonstrate models that contribute novel insights, rather than learning the most likely next step in a set of actions a clinician will take. We trained deep learning models using only clinician-initiated, administrative data for 42.9 million admissions using three subsets of data: demographic data only, demographic data and information available at admission, and the previous data plus charges recorded during the first day of admission. Models trained on charges during the first day of admission achieve performance close to published full EMR-based benchmarks for inpatient outcomes: inhospital mortality (0.89 AUC), prolonged length of stay (0.82 AUC), and 30-day readmission rate (0.71 AUC). Similar performance between models trained with only clinician-initiated data and those trained with full EMR data purporting to include information about patient state and physiology should raise concern in the deployment of these models. Furthermore, these models exhibited significant declines in performance when evaluated over only myocardial infarction (MI) patients relative to models trained over MI patients alone, highlighting the importance of physician diagnosis in the prognostic performance of these models. These results provide a benchmark for predictive accuracy trained only on prior clinical actions and indicate that models with similar performance may derive their signal by looking over clinician’s shoulders—using clinical behavior as the expression of preexisting intuition and suspicion to generate a prediction. For models to guide clinicians in individual decisions, performance exceeding these benchmarks is necessary. |
format | Online Article Text |
id | pubmed-8010071 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | Nature Publishing Group UK |
record_format | MEDLINE/PubMed |
spelling | pubmed-80100712021-04-16 Machine learning for patient risk stratification: standing on, or looking over, the shoulders of clinicians? Beaulieu-Jones, Brett K. Yuan, William Brat, Gabriel A. Beam, Andrew L. Weber, Griffin Ruffin, Marshall Kohane, Isaac S. NPJ Digit Med Article Machine learning can help clinicians to make individualized patient predictions only if researchers demonstrate models that contribute novel insights, rather than learning the most likely next step in a set of actions a clinician will take. We trained deep learning models using only clinician-initiated, administrative data for 42.9 million admissions using three subsets of data: demographic data only, demographic data and information available at admission, and the previous data plus charges recorded during the first day of admission. Models trained on charges during the first day of admission achieve performance close to published full EMR-based benchmarks for inpatient outcomes: inhospital mortality (0.89 AUC), prolonged length of stay (0.82 AUC), and 30-day readmission rate (0.71 AUC). Similar performance between models trained with only clinician-initiated data and those trained with full EMR data purporting to include information about patient state and physiology should raise concern in the deployment of these models. Furthermore, these models exhibited significant declines in performance when evaluated over only myocardial infarction (MI) patients relative to models trained over MI patients alone, highlighting the importance of physician diagnosis in the prognostic performance of these models. These results provide a benchmark for predictive accuracy trained only on prior clinical actions and indicate that models with similar performance may derive their signal by looking over clinician’s shoulders—using clinical behavior as the expression of preexisting intuition and suspicion to generate a prediction. For models to guide clinicians in individual decisions, performance exceeding these benchmarks is necessary. Nature Publishing Group UK 2021-03-30 /pmc/articles/PMC8010071/ /pubmed/33785839 http://dx.doi.org/10.1038/s41746-021-00426-3 Text en © The Author(s) 2021 Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/. |
spellingShingle | Article Beaulieu-Jones, Brett K. Yuan, William Brat, Gabriel A. Beam, Andrew L. Weber, Griffin Ruffin, Marshall Kohane, Isaac S. Machine learning for patient risk stratification: standing on, or looking over, the shoulders of clinicians? |
title | Machine learning for patient risk stratification: standing on, or looking over, the shoulders of clinicians? |
title_full | Machine learning for patient risk stratification: standing on, or looking over, the shoulders of clinicians? |
title_fullStr | Machine learning for patient risk stratification: standing on, or looking over, the shoulders of clinicians? |
title_full_unstemmed | Machine learning for patient risk stratification: standing on, or looking over, the shoulders of clinicians? |
title_short | Machine learning for patient risk stratification: standing on, or looking over, the shoulders of clinicians? |
title_sort | machine learning for patient risk stratification: standing on, or looking over, the shoulders of clinicians? |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8010071/ https://www.ncbi.nlm.nih.gov/pubmed/33785839 http://dx.doi.org/10.1038/s41746-021-00426-3 |
work_keys_str_mv | AT beaulieujonesbrettk machinelearningforpatientriskstratificationstandingonorlookingovertheshouldersofclinicians AT yuanwilliam machinelearningforpatientriskstratificationstandingonorlookingovertheshouldersofclinicians AT bratgabriela machinelearningforpatientriskstratificationstandingonorlookingovertheshouldersofclinicians AT beamandrewl machinelearningforpatientriskstratificationstandingonorlookingovertheshouldersofclinicians AT webergriffin machinelearningforpatientriskstratificationstandingonorlookingovertheshouldersofclinicians AT ruffinmarshall machinelearningforpatientriskstratificationstandingonorlookingovertheshouldersofclinicians AT kohaneisaacs machinelearningforpatientriskstratificationstandingonorlookingovertheshouldersofclinicians |