Cargando…

Interpretable machine learning prediction of all-cause mortality

BACKGROUND: Unlike linear models which are traditionally used to study all-cause mortality, complex machine learning models can capture non-linear interrelations and provide opportunities to identify unexplored risk factors. Explainable artificial intelligence can improve prediction accuracy over li...

Descripción completa

Detalles Bibliográficos
Autores principales: Qiu, Wei, Chen, Hugh, Dincer, Ayse Berceste, Lundberg, Scott, Kaeberlein, Matt, Lee, Su-In
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Nature Publishing Group UK 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9530124/
https://www.ncbi.nlm.nih.gov/pubmed/36204043
http://dx.doi.org/10.1038/s43856-022-00180-x
_version_ 1784801608420294656
author Qiu, Wei
Chen, Hugh
Dincer, Ayse Berceste
Lundberg, Scott
Kaeberlein, Matt
Lee, Su-In
author_facet Qiu, Wei
Chen, Hugh
Dincer, Ayse Berceste
Lundberg, Scott
Kaeberlein, Matt
Lee, Su-In
author_sort Qiu, Wei
collection PubMed
description BACKGROUND: Unlike linear models which are traditionally used to study all-cause mortality, complex machine learning models can capture non-linear interrelations and provide opportunities to identify unexplored risk factors. Explainable artificial intelligence can improve prediction accuracy over linear models and reveal great insights into outcomes like mortality. This paper comprehensively analyzes all-cause mortality by explaining complex machine learning models. METHODS: We propose the IMPACT framework that uses XAI technique to explain a state-of-the-art tree ensemble mortality prediction model. We apply IMPACT to understand all-cause mortality for 1-, 3-, 5-, and 10-year follow-up times within the NHANES dataset, which contains 47,261 samples and 151 features. RESULTS: We show that IMPACT models achieve higher accuracy than linear models and neural networks. Using IMPACT, we identify several overlooked risk factors and interaction effects. Furthermore, we identify relationships between laboratory features and mortality that may suggest adjusting established reference intervals. Finally, we develop highly accurate, efficient and interpretable mortality risk scores that can be used by medical professionals and individuals without medical expertise. We ensure generalizability by performing temporal validation of the mortality risk scores and external validation of important findings with the UK Biobank dataset. CONCLUSIONS: IMPACT’s unique strength is the explainable prediction, which provides insights into the complex, non-linear relationships between mortality and features, while maintaining high accuracy. Our explainable risk scores could help individuals improve self-awareness of their health status and help clinicians identify patients with high risk. IMPACT takes a consequential step towards bringing contemporary developments in XAI to epidemiology.
format Online
Article
Text
id pubmed-9530124
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Nature Publishing Group UK
record_format MEDLINE/PubMed
spelling pubmed-95301242022-10-05 Interpretable machine learning prediction of all-cause mortality Qiu, Wei Chen, Hugh Dincer, Ayse Berceste Lundberg, Scott Kaeberlein, Matt Lee, Su-In Commun Med (Lond) Article BACKGROUND: Unlike linear models which are traditionally used to study all-cause mortality, complex machine learning models can capture non-linear interrelations and provide opportunities to identify unexplored risk factors. Explainable artificial intelligence can improve prediction accuracy over linear models and reveal great insights into outcomes like mortality. This paper comprehensively analyzes all-cause mortality by explaining complex machine learning models. METHODS: We propose the IMPACT framework that uses XAI technique to explain a state-of-the-art tree ensemble mortality prediction model. We apply IMPACT to understand all-cause mortality for 1-, 3-, 5-, and 10-year follow-up times within the NHANES dataset, which contains 47,261 samples and 151 features. RESULTS: We show that IMPACT models achieve higher accuracy than linear models and neural networks. Using IMPACT, we identify several overlooked risk factors and interaction effects. Furthermore, we identify relationships between laboratory features and mortality that may suggest adjusting established reference intervals. Finally, we develop highly accurate, efficient and interpretable mortality risk scores that can be used by medical professionals and individuals without medical expertise. We ensure generalizability by performing temporal validation of the mortality risk scores and external validation of important findings with the UK Biobank dataset. CONCLUSIONS: IMPACT’s unique strength is the explainable prediction, which provides insights into the complex, non-linear relationships between mortality and features, while maintaining high accuracy. Our explainable risk scores could help individuals improve self-awareness of their health status and help clinicians identify patients with high risk. IMPACT takes a consequential step towards bringing contemporary developments in XAI to epidemiology. Nature Publishing Group UK 2022-10-03 /pmc/articles/PMC9530124/ /pubmed/36204043 http://dx.doi.org/10.1038/s43856-022-00180-x Text en © The Author(s) 2022 https://creativecommons.org/licenses/by/4.0/Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) .
spellingShingle Article
Qiu, Wei
Chen, Hugh
Dincer, Ayse Berceste
Lundberg, Scott
Kaeberlein, Matt
Lee, Su-In
Interpretable machine learning prediction of all-cause mortality
title Interpretable machine learning prediction of all-cause mortality
title_full Interpretable machine learning prediction of all-cause mortality
title_fullStr Interpretable machine learning prediction of all-cause mortality
title_full_unstemmed Interpretable machine learning prediction of all-cause mortality
title_short Interpretable machine learning prediction of all-cause mortality
title_sort interpretable machine learning prediction of all-cause mortality
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9530124/
https://www.ncbi.nlm.nih.gov/pubmed/36204043
http://dx.doi.org/10.1038/s43856-022-00180-x
work_keys_str_mv AT qiuwei interpretablemachinelearningpredictionofallcausemortality
AT chenhugh interpretablemachinelearningpredictionofallcausemortality
AT dincerayseberceste interpretablemachinelearningpredictionofallcausemortality
AT lundbergscott interpretablemachinelearningpredictionofallcausemortality
AT kaeberleinmatt interpretablemachinelearningpredictionofallcausemortality
AT leesuin interpretablemachinelearningpredictionofallcausemortality