Cargando…

Shapley variable importance cloud for interpretable machine learning

Interpretable machine learning has been focusing on explaining final models that optimize performance. The state-of-the-art Shapley additive explanations (SHAP) locally explains the variable impact on individual predictions and has recently been extended to provide global assessments across the data...

Descripción completa

Detalles Bibliográficos
Autores principales: Ning, Yilin, Ong, Marcus Eng Hock, Chakraborty, Bibhas, Goldstein, Benjamin Alan, Ting, Daniel Shu Wei, Vaughan, Roger, Liu, Nan
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Elsevier 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9023900/
https://www.ncbi.nlm.nih.gov/pubmed/35465224
http://dx.doi.org/10.1016/j.patter.2022.100452
_version_ 1784690442019799040
author Ning, Yilin
Ong, Marcus Eng Hock
Chakraborty, Bibhas
Goldstein, Benjamin Alan
Ting, Daniel Shu Wei
Vaughan, Roger
Liu, Nan
author_facet Ning, Yilin
Ong, Marcus Eng Hock
Chakraborty, Bibhas
Goldstein, Benjamin Alan
Ting, Daniel Shu Wei
Vaughan, Roger
Liu, Nan
author_sort Ning, Yilin
collection PubMed
description Interpretable machine learning has been focusing on explaining final models that optimize performance. The state-of-the-art Shapley additive explanations (SHAP) locally explains the variable impact on individual predictions and has recently been extended to provide global assessments across the dataset. Our work further extends “global” assessments to a set of models that are “good enough” and are practically as relevant as the final model to a prediction task. The resulting Shapley variable importance cloud consists of Shapley-based importance measures from each good model and pools information across models to provide an overall importance measure, with uncertainty explicitly quantified to support formal statistical inference. We developed visualizations to highlight the uncertainty and to illustrate its implications to practical inference. Building on a common theoretical basis, our method seamlessly complements the widely adopted SHAP assessments of a single final model to avoid biased inference, which we demonstrate in two experiments using recidivism prediction data and clinical data.
format Online
Article
Text
id pubmed-9023900
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Elsevier
record_format MEDLINE/PubMed
spelling pubmed-90239002022-04-23 Shapley variable importance cloud for interpretable machine learning Ning, Yilin Ong, Marcus Eng Hock Chakraborty, Bibhas Goldstein, Benjamin Alan Ting, Daniel Shu Wei Vaughan, Roger Liu, Nan Patterns (N Y) Article Interpretable machine learning has been focusing on explaining final models that optimize performance. The state-of-the-art Shapley additive explanations (SHAP) locally explains the variable impact on individual predictions and has recently been extended to provide global assessments across the dataset. Our work further extends “global” assessments to a set of models that are “good enough” and are practically as relevant as the final model to a prediction task. The resulting Shapley variable importance cloud consists of Shapley-based importance measures from each good model and pools information across models to provide an overall importance measure, with uncertainty explicitly quantified to support formal statistical inference. We developed visualizations to highlight the uncertainty and to illustrate its implications to practical inference. Building on a common theoretical basis, our method seamlessly complements the widely adopted SHAP assessments of a single final model to avoid biased inference, which we demonstrate in two experiments using recidivism prediction data and clinical data. Elsevier 2022-02-22 /pmc/articles/PMC9023900/ /pubmed/35465224 http://dx.doi.org/10.1016/j.patter.2022.100452 Text en © 2022 The Author(s) https://creativecommons.org/licenses/by/4.0/This is an open access article under the CC BY license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Ning, Yilin
Ong, Marcus Eng Hock
Chakraborty, Bibhas
Goldstein, Benjamin Alan
Ting, Daniel Shu Wei
Vaughan, Roger
Liu, Nan
Shapley variable importance cloud for interpretable machine learning
title Shapley variable importance cloud for interpretable machine learning
title_full Shapley variable importance cloud for interpretable machine learning
title_fullStr Shapley variable importance cloud for interpretable machine learning
title_full_unstemmed Shapley variable importance cloud for interpretable machine learning
title_short Shapley variable importance cloud for interpretable machine learning
title_sort shapley variable importance cloud for interpretable machine learning
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9023900/
https://www.ncbi.nlm.nih.gov/pubmed/35465224
http://dx.doi.org/10.1016/j.patter.2022.100452
work_keys_str_mv AT ningyilin shapleyvariableimportancecloudforinterpretablemachinelearning
AT ongmarcusenghock shapleyvariableimportancecloudforinterpretablemachinelearning
AT chakrabortybibhas shapleyvariableimportancecloudforinterpretablemachinelearning
AT goldsteinbenjaminalan shapleyvariableimportancecloudforinterpretablemachinelearning
AT tingdanielshuwei shapleyvariableimportancecloudforinterpretablemachinelearning
AT vaughanroger shapleyvariableimportancecloudforinterpretablemachinelearning
AT liunan shapleyvariableimportancecloudforinterpretablemachinelearning