Cargando…

Validation of parsimonious prognostic models for patients infected with COVID-19

OBJECTIVES: Predictive studies play important roles in the development of models informing care for patients with COVID-19. Our concern is that studies producing ill-performing models may lead to inappropriate clinical decision-making. Thus, our objective is to summarise and characterise performance...

Descripción completa

Detalles Bibliográficos
Autores principales: Harish, Keerthi, Zhang, Ben, Stella, Peter, Hauck, Kevin, Moussa, Marwa M, Adler, Nicole M, Horwitz, Leora I, Aphinyanaphongs, Yindalon
Formato: Online Artículo Texto
Lenguaje:English
Publicado: BMJ Publishing Group 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8421114/
https://www.ncbi.nlm.nih.gov/pubmed/34479962
http://dx.doi.org/10.1136/bmjhci-2020-100267
_version_ 1783749012845756416
author Harish, Keerthi
Zhang, Ben
Stella, Peter
Hauck, Kevin
Moussa, Marwa M
Adler, Nicole M
Horwitz, Leora I
Aphinyanaphongs, Yindalon
author_facet Harish, Keerthi
Zhang, Ben
Stella, Peter
Hauck, Kevin
Moussa, Marwa M
Adler, Nicole M
Horwitz, Leora I
Aphinyanaphongs, Yindalon
author_sort Harish, Keerthi
collection PubMed
description OBJECTIVES: Predictive studies play important roles in the development of models informing care for patients with COVID-19. Our concern is that studies producing ill-performing models may lead to inappropriate clinical decision-making. Thus, our objective is to summarise and characterise performance of prognostic models for COVID-19 on external data. METHODS: We performed a validation of parsimonious prognostic models for patients with COVID-19 from a literature search for published and preprint articles. Ten models meeting inclusion criteria were either (a) externally validated with our data against the model variables and weights or (b) rebuilt using original features if no weights were provided. Nine studies had internally or externally validated models on cohorts of between 18 and 320 inpatients with COVID-19. One model used cross-validation. Our external validation cohort consisted of 4444 patients with COVID-19 hospitalised between 1 March and 27 May 2020. RESULTS: Most models failed validation when applied to our institution’s data. Included studies reported an average validation area under the receiver–operator curve (AUROC) of 0.828. Models applied with reported features averaged an AUROC of 0.66 when validated on our data. Models rebuilt with the same features averaged an AUROC of 0.755 when validated on our data. In both cases, models did not validate against their studies’ reported AUROC values. DISCUSSION: Published and preprint prognostic models for patients infected with COVID-19 performed substantially worse when applied to external data. Further inquiry is required to elucidate mechanisms underlying performance deviations. CONCLUSIONS: Clinicians should employ caution when applying models for clinical prediction without careful validation on local data.
format Online
Article
Text
id pubmed-8421114
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher BMJ Publishing Group
record_format MEDLINE/PubMed
spelling pubmed-84211142021-09-07 Validation of parsimonious prognostic models for patients infected with COVID-19 Harish, Keerthi Zhang, Ben Stella, Peter Hauck, Kevin Moussa, Marwa M Adler, Nicole M Horwitz, Leora I Aphinyanaphongs, Yindalon BMJ Health Care Inform Original Research OBJECTIVES: Predictive studies play important roles in the development of models informing care for patients with COVID-19. Our concern is that studies producing ill-performing models may lead to inappropriate clinical decision-making. Thus, our objective is to summarise and characterise performance of prognostic models for COVID-19 on external data. METHODS: We performed a validation of parsimonious prognostic models for patients with COVID-19 from a literature search for published and preprint articles. Ten models meeting inclusion criteria were either (a) externally validated with our data against the model variables and weights or (b) rebuilt using original features if no weights were provided. Nine studies had internally or externally validated models on cohorts of between 18 and 320 inpatients with COVID-19. One model used cross-validation. Our external validation cohort consisted of 4444 patients with COVID-19 hospitalised between 1 March and 27 May 2020. RESULTS: Most models failed validation when applied to our institution’s data. Included studies reported an average validation area under the receiver–operator curve (AUROC) of 0.828. Models applied with reported features averaged an AUROC of 0.66 when validated on our data. Models rebuilt with the same features averaged an AUROC of 0.755 when validated on our data. In both cases, models did not validate against their studies’ reported AUROC values. DISCUSSION: Published and preprint prognostic models for patients infected with COVID-19 performed substantially worse when applied to external data. Further inquiry is required to elucidate mechanisms underlying performance deviations. CONCLUSIONS: Clinicians should employ caution when applying models for clinical prediction without careful validation on local data. BMJ Publishing Group 2021-09-03 /pmc/articles/PMC8421114/ /pubmed/34479962 http://dx.doi.org/10.1136/bmjhci-2020-100267 Text en © Author(s) (or their employer(s)) 2021. Re-use permitted under CC BY-NC. No commercial re-use. See rights and permissions. Published by BMJ. https://creativecommons.org/licenses/by-nc/4.0/This is an open access article distributed in accordance with the Creative Commons Attribution Non Commercial (CC BY-NC 4.0) license, which permits others to distribute, remix, adapt, build upon this work non-commercially, and license their derivative works on different terms, provided the original work is properly cited, appropriate credit is given, any changes made indicated, and the use is non-commercial. See: http://creativecommons.org/licenses/by-nc/4.0/ (https://creativecommons.org/licenses/by-nc/4.0/) .
spellingShingle Original Research
Harish, Keerthi
Zhang, Ben
Stella, Peter
Hauck, Kevin
Moussa, Marwa M
Adler, Nicole M
Horwitz, Leora I
Aphinyanaphongs, Yindalon
Validation of parsimonious prognostic models for patients infected with COVID-19
title Validation of parsimonious prognostic models for patients infected with COVID-19
title_full Validation of parsimonious prognostic models for patients infected with COVID-19
title_fullStr Validation of parsimonious prognostic models for patients infected with COVID-19
title_full_unstemmed Validation of parsimonious prognostic models for patients infected with COVID-19
title_short Validation of parsimonious prognostic models for patients infected with COVID-19
title_sort validation of parsimonious prognostic models for patients infected with covid-19
topic Original Research
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8421114/
https://www.ncbi.nlm.nih.gov/pubmed/34479962
http://dx.doi.org/10.1136/bmjhci-2020-100267
work_keys_str_mv AT harishkeerthi validationofparsimoniousprognosticmodelsforpatientsinfectedwithcovid19
AT zhangben validationofparsimoniousprognosticmodelsforpatientsinfectedwithcovid19
AT stellapeter validationofparsimoniousprognosticmodelsforpatientsinfectedwithcovid19
AT hauckkevin validationofparsimoniousprognosticmodelsforpatientsinfectedwithcovid19
AT moussamarwam validationofparsimoniousprognosticmodelsforpatientsinfectedwithcovid19
AT adlernicolem validationofparsimoniousprognosticmodelsforpatientsinfectedwithcovid19
AT horwitzleorai validationofparsimoniousprognosticmodelsforpatientsinfectedwithcovid19
AT aphinyanaphongsyindalon validationofparsimoniousprognosticmodelsforpatientsinfectedwithcovid19