Cargando…
Development and validation of an interpretable 3 day intensive care unit readmission prediction model using explainable boosting machines
BACKGROUND: Intensive care unit (ICU) readmissions are associated with mortality and poor outcomes. To improve discharge decisions, machine learning (ML) could help to identify patients at risk of ICU readmission. However, as many models are black boxes, dangerous properties may remain unnoticed. Wi...
Autores principales: | , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9445989/ https://www.ncbi.nlm.nih.gov/pubmed/36082270 http://dx.doi.org/10.3389/fmed.2022.960296 |
_version_ | 1784783544801820672 |
---|---|
author | Hegselmann, Stefan Ertmer, Christian Volkert, Thomas Gottschalk, Antje Dugas, Martin Varghese, Julian |
author_facet | Hegselmann, Stefan Ertmer, Christian Volkert, Thomas Gottschalk, Antje Dugas, Martin Varghese, Julian |
author_sort | Hegselmann, Stefan |
collection | PubMed |
description | BACKGROUND: Intensive care unit (ICU) readmissions are associated with mortality and poor outcomes. To improve discharge decisions, machine learning (ML) could help to identify patients at risk of ICU readmission. However, as many models are black boxes, dangerous properties may remain unnoticed. Widely used post hoc explanation methods also have inherent limitations. Few studies are evaluating inherently interpretable ML models for health care and involve clinicians in inspecting the trained model. METHODS: An inherently interpretable model for the prediction of 3 day ICU readmission was developed. We used explainable boosting machines that learn modular risk functions and which have already been shown to be suitable for the health care domain. We created a retrospective cohort of 15,589 ICU stays and 169 variables collected between 2006 and 2019 from the University Hospital Münster. A team of physicians inspected the model, checked the plausibility of each risk function, and removed problematic ones. We collected qualitative feedback during this process and analyzed the reasons for removing risk functions. The performance of the final explainable boosting machine was compared with a validated clinical score and three commonly used ML models. External validation was performed on the widely used Medical Information Mart for Intensive Care version IV database. RESULTS: The developed explainable boosting machine used 67 features and showed an area under the precision-recall curve of 0.119 ± 0.020 and an area under the receiver operating characteristic curve of 0.680 ± 0.025. It performed on par with state-of-the-art gradient boosting machines (0.123 ± 0.016, 0.665 ± 0.036) and outperformed the Simplified Acute Physiology Score II (0.084 ± 0.025, 0.607 ± 0.019), logistic regression (0.092 ± 0.026, 0.587 ± 0.016), and recurrent neural networks (0.095 ± 0.008, 0.594 ± 0.027). External validation confirmed that explainable boosting machines (0.221 ± 0.023, 0.760 ± 0.010) performed similarly to gradient boosting machines (0.232 ± 0.029, 0.772 ± 0.018). Evaluation of the model inspection showed that explainable boosting machines can be useful to detect and remove problematic risk functions. CONCLUSIONS: We developed an inherently interpretable ML model for 3 day ICU readmission prediction that reached the state-of-the-art performance of black box models. Our results suggest that for low- to medium-dimensional datasets that are common in health care, it is feasible to develop ML models that allow a high level of human control without sacrificing performance. |
format | Online Article Text |
id | pubmed-9445989 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-94459892022-09-07 Development and validation of an interpretable 3 day intensive care unit readmission prediction model using explainable boosting machines Hegselmann, Stefan Ertmer, Christian Volkert, Thomas Gottschalk, Antje Dugas, Martin Varghese, Julian Front Med (Lausanne) Medicine BACKGROUND: Intensive care unit (ICU) readmissions are associated with mortality and poor outcomes. To improve discharge decisions, machine learning (ML) could help to identify patients at risk of ICU readmission. However, as many models are black boxes, dangerous properties may remain unnoticed. Widely used post hoc explanation methods also have inherent limitations. Few studies are evaluating inherently interpretable ML models for health care and involve clinicians in inspecting the trained model. METHODS: An inherently interpretable model for the prediction of 3 day ICU readmission was developed. We used explainable boosting machines that learn modular risk functions and which have already been shown to be suitable for the health care domain. We created a retrospective cohort of 15,589 ICU stays and 169 variables collected between 2006 and 2019 from the University Hospital Münster. A team of physicians inspected the model, checked the plausibility of each risk function, and removed problematic ones. We collected qualitative feedback during this process and analyzed the reasons for removing risk functions. The performance of the final explainable boosting machine was compared with a validated clinical score and three commonly used ML models. External validation was performed on the widely used Medical Information Mart for Intensive Care version IV database. RESULTS: The developed explainable boosting machine used 67 features and showed an area under the precision-recall curve of 0.119 ± 0.020 and an area under the receiver operating characteristic curve of 0.680 ± 0.025. It performed on par with state-of-the-art gradient boosting machines (0.123 ± 0.016, 0.665 ± 0.036) and outperformed the Simplified Acute Physiology Score II (0.084 ± 0.025, 0.607 ± 0.019), logistic regression (0.092 ± 0.026, 0.587 ± 0.016), and recurrent neural networks (0.095 ± 0.008, 0.594 ± 0.027). External validation confirmed that explainable boosting machines (0.221 ± 0.023, 0.760 ± 0.010) performed similarly to gradient boosting machines (0.232 ± 0.029, 0.772 ± 0.018). Evaluation of the model inspection showed that explainable boosting machines can be useful to detect and remove problematic risk functions. CONCLUSIONS: We developed an inherently interpretable ML model for 3 day ICU readmission prediction that reached the state-of-the-art performance of black box models. Our results suggest that for low- to medium-dimensional datasets that are common in health care, it is feasible to develop ML models that allow a high level of human control without sacrificing performance. Frontiers Media S.A. 2022-08-23 /pmc/articles/PMC9445989/ /pubmed/36082270 http://dx.doi.org/10.3389/fmed.2022.960296 Text en Copyright © 2022 Hegselmann, Ertmer, Volkert, Gottschalk, Dugas and Varghese. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Medicine Hegselmann, Stefan Ertmer, Christian Volkert, Thomas Gottschalk, Antje Dugas, Martin Varghese, Julian Development and validation of an interpretable 3 day intensive care unit readmission prediction model using explainable boosting machines |
title | Development and validation of an interpretable 3 day intensive care unit readmission prediction model using explainable boosting machines |
title_full | Development and validation of an interpretable 3 day intensive care unit readmission prediction model using explainable boosting machines |
title_fullStr | Development and validation of an interpretable 3 day intensive care unit readmission prediction model using explainable boosting machines |
title_full_unstemmed | Development and validation of an interpretable 3 day intensive care unit readmission prediction model using explainable boosting machines |
title_short | Development and validation of an interpretable 3 day intensive care unit readmission prediction model using explainable boosting machines |
title_sort | development and validation of an interpretable 3 day intensive care unit readmission prediction model using explainable boosting machines |
topic | Medicine |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9445989/ https://www.ncbi.nlm.nih.gov/pubmed/36082270 http://dx.doi.org/10.3389/fmed.2022.960296 |
work_keys_str_mv | AT hegselmannstefan developmentandvalidationofaninterpretable3dayintensivecareunitreadmissionpredictionmodelusingexplainableboostingmachines AT ertmerchristian developmentandvalidationofaninterpretable3dayintensivecareunitreadmissionpredictionmodelusingexplainableboostingmachines AT volkertthomas developmentandvalidationofaninterpretable3dayintensivecareunitreadmissionpredictionmodelusingexplainableboostingmachines AT gottschalkantje developmentandvalidationofaninterpretable3dayintensivecareunitreadmissionpredictionmodelusingexplainableboostingmachines AT dugasmartin developmentandvalidationofaninterpretable3dayintensivecareunitreadmissionpredictionmodelusingexplainableboostingmachines AT varghesejulian developmentandvalidationofaninterpretable3dayintensivecareunitreadmissionpredictionmodelusingexplainableboostingmachines |