Cargando…

On the interpretability of machine learning-based model for predicting hypertension

BACKGROUND: Although complex machine learning models are commonly outperforming the traditional simple interpretable models, clinicians find it hard to understand and trust these complex models due to the lack of intuition and explanation of their predictions. The aim of this study to demonstrate th...

Descripción completa

Detalles Bibliográficos
Autores principales: Elshawi, Radwa, Al-Mallah, Mouaz H., Sakr, Sherif
Formato: Online Artículo Texto
Lenguaje:English
Publicado: BioMed Central 2019
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6664803/
https://www.ncbi.nlm.nih.gov/pubmed/31357998
http://dx.doi.org/10.1186/s12911-019-0874-0
_version_ 1783439960736530432
author Elshawi, Radwa
Al-Mallah, Mouaz H.
Sakr, Sherif
author_facet Elshawi, Radwa
Al-Mallah, Mouaz H.
Sakr, Sherif
author_sort Elshawi, Radwa
collection PubMed
description BACKGROUND: Although complex machine learning models are commonly outperforming the traditional simple interpretable models, clinicians find it hard to understand and trust these complex models due to the lack of intuition and explanation of their predictions. The aim of this study to demonstrate the utility of various model-agnostic explanation techniques of machine learning models with a case study for analyzing the outcomes of the machine learning random forest model for predicting the individuals at risk of developing hypertension based on cardiorespiratory fitness data. METHODS: The dataset used in this study contains information of 23,095 patients who underwent clinician-referred exercise treadmill stress testing at Henry Ford Health Systems between 1991 and 2009 and had a complete 10-year follow-up. Five global interpretability techniques (Feature Importance, Partial Dependence Plot, Individual Conditional Expectation, Feature Interaction, Global Surrogate Models) and two local interpretability techniques (Local Surrogate Models, Shapley Value) have been applied to present the role of the interpretability techniques on assisting the clinical staff to get better understanding and more trust of the outcomes of the machine learning-based predictions. RESULTS: Several experiments have been conducted and reported. The results show that different interpretability techniques can shed light on different insights on the model behavior where global interpretations can enable clinicians to understand the entire conditional distribution modeled by the trained response function. In contrast, local interpretations promote the understanding of small parts of the conditional distribution for specific instances. CONCLUSIONS: Various interpretability techniques can vary in their explanations for the behavior of the machine learning model. The global interpretability techniques have the advantage that it can generalize over the entire population while local interpretability techniques focus on giving explanations at the level of instances. Both methods can be equally valid depending on the application need. Both methods are effective methods for assisting clinicians on the medical decision process, however, the clinicians will always remain to hold the final say on accepting or rejecting the outcome of the machine learning models and their explanations based on their domain expertise.
format Online
Article
Text
id pubmed-6664803
institution National Center for Biotechnology Information
language English
publishDate 2019
publisher BioMed Central
record_format MEDLINE/PubMed
spelling pubmed-66648032019-08-05 On the interpretability of machine learning-based model for predicting hypertension Elshawi, Radwa Al-Mallah, Mouaz H. Sakr, Sherif BMC Med Inform Decis Mak Research Article BACKGROUND: Although complex machine learning models are commonly outperforming the traditional simple interpretable models, clinicians find it hard to understand and trust these complex models due to the lack of intuition and explanation of their predictions. The aim of this study to demonstrate the utility of various model-agnostic explanation techniques of machine learning models with a case study for analyzing the outcomes of the machine learning random forest model for predicting the individuals at risk of developing hypertension based on cardiorespiratory fitness data. METHODS: The dataset used in this study contains information of 23,095 patients who underwent clinician-referred exercise treadmill stress testing at Henry Ford Health Systems between 1991 and 2009 and had a complete 10-year follow-up. Five global interpretability techniques (Feature Importance, Partial Dependence Plot, Individual Conditional Expectation, Feature Interaction, Global Surrogate Models) and two local interpretability techniques (Local Surrogate Models, Shapley Value) have been applied to present the role of the interpretability techniques on assisting the clinical staff to get better understanding and more trust of the outcomes of the machine learning-based predictions. RESULTS: Several experiments have been conducted and reported. The results show that different interpretability techniques can shed light on different insights on the model behavior where global interpretations can enable clinicians to understand the entire conditional distribution modeled by the trained response function. In contrast, local interpretations promote the understanding of small parts of the conditional distribution for specific instances. CONCLUSIONS: Various interpretability techniques can vary in their explanations for the behavior of the machine learning model. The global interpretability techniques have the advantage that it can generalize over the entire population while local interpretability techniques focus on giving explanations at the level of instances. Both methods can be equally valid depending on the application need. Both methods are effective methods for assisting clinicians on the medical decision process, however, the clinicians will always remain to hold the final say on accepting or rejecting the outcome of the machine learning models and their explanations based on their domain expertise. BioMed Central 2019-07-29 /pmc/articles/PMC6664803/ /pubmed/31357998 http://dx.doi.org/10.1186/s12911-019-0874-0 Text en © The Author(s). 2019 Open AccessThis article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.
spellingShingle Research Article
Elshawi, Radwa
Al-Mallah, Mouaz H.
Sakr, Sherif
On the interpretability of machine learning-based model for predicting hypertension
title On the interpretability of machine learning-based model for predicting hypertension
title_full On the interpretability of machine learning-based model for predicting hypertension
title_fullStr On the interpretability of machine learning-based model for predicting hypertension
title_full_unstemmed On the interpretability of machine learning-based model for predicting hypertension
title_short On the interpretability of machine learning-based model for predicting hypertension
title_sort on the interpretability of machine learning-based model for predicting hypertension
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6664803/
https://www.ncbi.nlm.nih.gov/pubmed/31357998
http://dx.doi.org/10.1186/s12911-019-0874-0
work_keys_str_mv AT elshawiradwa ontheinterpretabilityofmachinelearningbasedmodelforpredictinghypertension
AT almallahmouazh ontheinterpretabilityofmachinelearningbasedmodelforpredictinghypertension
AT sakrsherif ontheinterpretabilityofmachinelearningbasedmodelforpredictinghypertension