Cargando…

Development and validation of an interpretable 3 day intensive care unit readmission prediction model using explainable boosting machines

BACKGROUND: Intensive care unit (ICU) readmissions are associated with mortality and poor outcomes. To improve discharge decisions, machine learning (ML) could help to identify patients at risk of ICU readmission. However, as many models are black boxes, dangerous properties may remain unnoticed. Wi...

Descripción completa

Detalles Bibliográficos
Autores principales: Hegselmann, Stefan, Ertmer, Christian, Volkert, Thomas, Gottschalk, Antje, Dugas, Martin, Varghese, Julian
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9445989/
https://www.ncbi.nlm.nih.gov/pubmed/36082270
http://dx.doi.org/10.3389/fmed.2022.960296
_version_ 1784783544801820672
author Hegselmann, Stefan
Ertmer, Christian
Volkert, Thomas
Gottschalk, Antje
Dugas, Martin
Varghese, Julian
author_facet Hegselmann, Stefan
Ertmer, Christian
Volkert, Thomas
Gottschalk, Antje
Dugas, Martin
Varghese, Julian
author_sort Hegselmann, Stefan
collection PubMed
description BACKGROUND: Intensive care unit (ICU) readmissions are associated with mortality and poor outcomes. To improve discharge decisions, machine learning (ML) could help to identify patients at risk of ICU readmission. However, as many models are black boxes, dangerous properties may remain unnoticed. Widely used post hoc explanation methods also have inherent limitations. Few studies are evaluating inherently interpretable ML models for health care and involve clinicians in inspecting the trained model. METHODS: An inherently interpretable model for the prediction of 3 day ICU readmission was developed. We used explainable boosting machines that learn modular risk functions and which have already been shown to be suitable for the health care domain. We created a retrospective cohort of 15,589 ICU stays and 169 variables collected between 2006 and 2019 from the University Hospital Münster. A team of physicians inspected the model, checked the plausibility of each risk function, and removed problematic ones. We collected qualitative feedback during this process and analyzed the reasons for removing risk functions. The performance of the final explainable boosting machine was compared with a validated clinical score and three commonly used ML models. External validation was performed on the widely used Medical Information Mart for Intensive Care version IV database. RESULTS: The developed explainable boosting machine used 67 features and showed an area under the precision-recall curve of 0.119 ± 0.020 and an area under the receiver operating characteristic curve of 0.680 ± 0.025. It performed on par with state-of-the-art gradient boosting machines (0.123 ± 0.016, 0.665 ± 0.036) and outperformed the Simplified Acute Physiology Score II (0.084 ± 0.025, 0.607 ± 0.019), logistic regression (0.092 ± 0.026, 0.587 ± 0.016), and recurrent neural networks (0.095 ± 0.008, 0.594 ± 0.027). External validation confirmed that explainable boosting machines (0.221 ± 0.023, 0.760 ± 0.010) performed similarly to gradient boosting machines (0.232 ± 0.029, 0.772 ± 0.018). Evaluation of the model inspection showed that explainable boosting machines can be useful to detect and remove problematic risk functions. CONCLUSIONS: We developed an inherently interpretable ML model for 3 day ICU readmission prediction that reached the state-of-the-art performance of black box models. Our results suggest that for low- to medium-dimensional datasets that are common in health care, it is feasible to develop ML models that allow a high level of human control without sacrificing performance.
format Online
Article
Text
id pubmed-9445989
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-94459892022-09-07 Development and validation of an interpretable 3 day intensive care unit readmission prediction model using explainable boosting machines Hegselmann, Stefan Ertmer, Christian Volkert, Thomas Gottschalk, Antje Dugas, Martin Varghese, Julian Front Med (Lausanne) Medicine BACKGROUND: Intensive care unit (ICU) readmissions are associated with mortality and poor outcomes. To improve discharge decisions, machine learning (ML) could help to identify patients at risk of ICU readmission. However, as many models are black boxes, dangerous properties may remain unnoticed. Widely used post hoc explanation methods also have inherent limitations. Few studies are evaluating inherently interpretable ML models for health care and involve clinicians in inspecting the trained model. METHODS: An inherently interpretable model for the prediction of 3 day ICU readmission was developed. We used explainable boosting machines that learn modular risk functions and which have already been shown to be suitable for the health care domain. We created a retrospective cohort of 15,589 ICU stays and 169 variables collected between 2006 and 2019 from the University Hospital Münster. A team of physicians inspected the model, checked the plausibility of each risk function, and removed problematic ones. We collected qualitative feedback during this process and analyzed the reasons for removing risk functions. The performance of the final explainable boosting machine was compared with a validated clinical score and three commonly used ML models. External validation was performed on the widely used Medical Information Mart for Intensive Care version IV database. RESULTS: The developed explainable boosting machine used 67 features and showed an area under the precision-recall curve of 0.119 ± 0.020 and an area under the receiver operating characteristic curve of 0.680 ± 0.025. It performed on par with state-of-the-art gradient boosting machines (0.123 ± 0.016, 0.665 ± 0.036) and outperformed the Simplified Acute Physiology Score II (0.084 ± 0.025, 0.607 ± 0.019), logistic regression (0.092 ± 0.026, 0.587 ± 0.016), and recurrent neural networks (0.095 ± 0.008, 0.594 ± 0.027). External validation confirmed that explainable boosting machines (0.221 ± 0.023, 0.760 ± 0.010) performed similarly to gradient boosting machines (0.232 ± 0.029, 0.772 ± 0.018). Evaluation of the model inspection showed that explainable boosting machines can be useful to detect and remove problematic risk functions. CONCLUSIONS: We developed an inherently interpretable ML model for 3 day ICU readmission prediction that reached the state-of-the-art performance of black box models. Our results suggest that for low- to medium-dimensional datasets that are common in health care, it is feasible to develop ML models that allow a high level of human control without sacrificing performance. Frontiers Media S.A. 2022-08-23 /pmc/articles/PMC9445989/ /pubmed/36082270 http://dx.doi.org/10.3389/fmed.2022.960296 Text en Copyright © 2022 Hegselmann, Ertmer, Volkert, Gottschalk, Dugas and Varghese. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Medicine
Hegselmann, Stefan
Ertmer, Christian
Volkert, Thomas
Gottschalk, Antje
Dugas, Martin
Varghese, Julian
Development and validation of an interpretable 3 day intensive care unit readmission prediction model using explainable boosting machines
title Development and validation of an interpretable 3 day intensive care unit readmission prediction model using explainable boosting machines
title_full Development and validation of an interpretable 3 day intensive care unit readmission prediction model using explainable boosting machines
title_fullStr Development and validation of an interpretable 3 day intensive care unit readmission prediction model using explainable boosting machines
title_full_unstemmed Development and validation of an interpretable 3 day intensive care unit readmission prediction model using explainable boosting machines
title_short Development and validation of an interpretable 3 day intensive care unit readmission prediction model using explainable boosting machines
title_sort development and validation of an interpretable 3 day intensive care unit readmission prediction model using explainable boosting machines
topic Medicine
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9445989/
https://www.ncbi.nlm.nih.gov/pubmed/36082270
http://dx.doi.org/10.3389/fmed.2022.960296
work_keys_str_mv AT hegselmannstefan developmentandvalidationofaninterpretable3dayintensivecareunitreadmissionpredictionmodelusingexplainableboostingmachines
AT ertmerchristian developmentandvalidationofaninterpretable3dayintensivecareunitreadmissionpredictionmodelusingexplainableboostingmachines
AT volkertthomas developmentandvalidationofaninterpretable3dayintensivecareunitreadmissionpredictionmodelusingexplainableboostingmachines
AT gottschalkantje developmentandvalidationofaninterpretable3dayintensivecareunitreadmissionpredictionmodelusingexplainableboostingmachines
AT dugasmartin developmentandvalidationofaninterpretable3dayintensivecareunitreadmissionpredictionmodelusingexplainableboostingmachines
AT varghesejulian developmentandvalidationofaninterpretable3dayintensivecareunitreadmissionpredictionmodelusingexplainableboostingmachines