Cargando…

Development and validation of an interpretable neural network for prediction of postoperative in-hospital mortality

While deep neural networks (DNNs) and other machine learning models often have higher accuracy than simpler models like logistic regression (LR), they are often considered to be “black box” models and this lack of interpretability and transparency is considered a challenge for clinical adoption. In...

Descripción completa

Detalles Bibliográficos
Autores principales: Lee, Christine K., Samad, Muntaha, Hofer, Ira, Cannesson, Maxime, Baldi, Pierre
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Nature Publishing Group UK 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7794438/
https://www.ncbi.nlm.nih.gov/pubmed/33420341
http://dx.doi.org/10.1038/s41746-020-00377-1
_version_ 1783634208675069952
author Lee, Christine K.
Samad, Muntaha
Hofer, Ira
Cannesson, Maxime
Baldi, Pierre
author_facet Lee, Christine K.
Samad, Muntaha
Hofer, Ira
Cannesson, Maxime
Baldi, Pierre
author_sort Lee, Christine K.
collection PubMed
description While deep neural networks (DNNs) and other machine learning models often have higher accuracy than simpler models like logistic regression (LR), they are often considered to be “black box” models and this lack of interpretability and transparency is considered a challenge for clinical adoption. In healthcare, intelligible models not only help clinicians to understand the problem and create more targeted action plans, but also help to gain the clinicians’ trust. One method of overcoming the limited interpretability of more complex models is to use Generalized Additive Models (GAMs). Standard GAMs simply model the target response as a sum of univariate models. Inspired by GAMs, the same idea can be applied to neural networks through an architecture referred to as Generalized Additive Models with Neural Networks (GAM-NNs). In this manuscript, we present the development and validation of a model applying the concept of GAM-NNs to allow for interpretability by visualizing the learned feature patterns related to risk of in-hospital mortality for patients undergoing surgery under general anesthesia. The data consists of 59,985 patients with a feature set of 46 features extracted at the end of surgery to which we added previously not included features: total anesthesia case time (1 feature); the time in minutes spent with mean arterial pressure (MAP) below 40, 45, 50, 55, 60, and 65 mmHg during surgery (6 features); and Healthcare Cost and Utilization Project (HCUP) Code Descriptions of the Primary current procedure terminology (CPT) codes (33 features) for a total of 86 features. All data were randomly split into 80% for training (n = 47,988) and 20% for testing (n = 11,997) prior to model development. Model performance was compared to a standard LR model using the same features as the GAM-NN. The data consisted of 59,985 surgical records, and the occurrence of in-hospital mortality was 0.81% in the training set and 0.72% in the testing set. The GAM-NN model with HCUP features had the highest area under the curve (AUC) 0.921 (0.895–0.95). Overall, both GAM-NN models had higher AUCs than LR models, however, had lower average precisions. The LR model without HCUP features had the highest average precision 0.217 (0.136–0.31). To assess the interpretability of the GAM-NNs, we then visualized the learned contributions of the GAM-NNs and compared against the learned contributions of the LRs for the models with HCUP features. Overall, we were able to demonstrate that our proposed generalized additive neural network (GAM-NN) architecture is able to (1) leverage a neural network’s ability to learn nonlinear patterns in the data, which is more clinically intuitive, (2) be interpreted easily, making it more clinically useful, and (3) maintain model performance as compared to previously published DNNs.
format Online
Article
Text
id pubmed-7794438
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher Nature Publishing Group UK
record_format MEDLINE/PubMed
spelling pubmed-77944382021-01-21 Development and validation of an interpretable neural network for prediction of postoperative in-hospital mortality Lee, Christine K. Samad, Muntaha Hofer, Ira Cannesson, Maxime Baldi, Pierre NPJ Digit Med Article While deep neural networks (DNNs) and other machine learning models often have higher accuracy than simpler models like logistic regression (LR), they are often considered to be “black box” models and this lack of interpretability and transparency is considered a challenge for clinical adoption. In healthcare, intelligible models not only help clinicians to understand the problem and create more targeted action plans, but also help to gain the clinicians’ trust. One method of overcoming the limited interpretability of more complex models is to use Generalized Additive Models (GAMs). Standard GAMs simply model the target response as a sum of univariate models. Inspired by GAMs, the same idea can be applied to neural networks through an architecture referred to as Generalized Additive Models with Neural Networks (GAM-NNs). In this manuscript, we present the development and validation of a model applying the concept of GAM-NNs to allow for interpretability by visualizing the learned feature patterns related to risk of in-hospital mortality for patients undergoing surgery under general anesthesia. The data consists of 59,985 patients with a feature set of 46 features extracted at the end of surgery to which we added previously not included features: total anesthesia case time (1 feature); the time in minutes spent with mean arterial pressure (MAP) below 40, 45, 50, 55, 60, and 65 mmHg during surgery (6 features); and Healthcare Cost and Utilization Project (HCUP) Code Descriptions of the Primary current procedure terminology (CPT) codes (33 features) for a total of 86 features. All data were randomly split into 80% for training (n = 47,988) and 20% for testing (n = 11,997) prior to model development. Model performance was compared to a standard LR model using the same features as the GAM-NN. The data consisted of 59,985 surgical records, and the occurrence of in-hospital mortality was 0.81% in the training set and 0.72% in the testing set. The GAM-NN model with HCUP features had the highest area under the curve (AUC) 0.921 (0.895–0.95). Overall, both GAM-NN models had higher AUCs than LR models, however, had lower average precisions. The LR model without HCUP features had the highest average precision 0.217 (0.136–0.31). To assess the interpretability of the GAM-NNs, we then visualized the learned contributions of the GAM-NNs and compared against the learned contributions of the LRs for the models with HCUP features. Overall, we were able to demonstrate that our proposed generalized additive neural network (GAM-NN) architecture is able to (1) leverage a neural network’s ability to learn nonlinear patterns in the data, which is more clinically intuitive, (2) be interpreted easily, making it more clinically useful, and (3) maintain model performance as compared to previously published DNNs. Nature Publishing Group UK 2021-01-08 /pmc/articles/PMC7794438/ /pubmed/33420341 http://dx.doi.org/10.1038/s41746-020-00377-1 Text en © The Author(s) 2021 Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.
spellingShingle Article
Lee, Christine K.
Samad, Muntaha
Hofer, Ira
Cannesson, Maxime
Baldi, Pierre
Development and validation of an interpretable neural network for prediction of postoperative in-hospital mortality
title Development and validation of an interpretable neural network for prediction of postoperative in-hospital mortality
title_full Development and validation of an interpretable neural network for prediction of postoperative in-hospital mortality
title_fullStr Development and validation of an interpretable neural network for prediction of postoperative in-hospital mortality
title_full_unstemmed Development and validation of an interpretable neural network for prediction of postoperative in-hospital mortality
title_short Development and validation of an interpretable neural network for prediction of postoperative in-hospital mortality
title_sort development and validation of an interpretable neural network for prediction of postoperative in-hospital mortality
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7794438/
https://www.ncbi.nlm.nih.gov/pubmed/33420341
http://dx.doi.org/10.1038/s41746-020-00377-1
work_keys_str_mv AT leechristinek developmentandvalidationofaninterpretableneuralnetworkforpredictionofpostoperativeinhospitalmortality
AT samadmuntaha developmentandvalidationofaninterpretableneuralnetworkforpredictionofpostoperativeinhospitalmortality
AT hoferira developmentandvalidationofaninterpretableneuralnetworkforpredictionofpostoperativeinhospitalmortality
AT cannessonmaxime developmentandvalidationofaninterpretableneuralnetworkforpredictionofpostoperativeinhospitalmortality
AT baldipierre developmentandvalidationofaninterpretableneuralnetworkforpredictionofpostoperativeinhospitalmortality