Cargando…
Shapley variable importance cloud for interpretable machine learning
Interpretable machine learning has been focusing on explaining final models that optimize performance. The state-of-the-art Shapley additive explanations (SHAP) locally explains the variable impact on individual predictions and has recently been extended to provide global assessments across the data...
Autores principales: | , , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Elsevier
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9023900/ https://www.ncbi.nlm.nih.gov/pubmed/35465224 http://dx.doi.org/10.1016/j.patter.2022.100452 |
_version_ | 1784690442019799040 |
---|---|
author | Ning, Yilin Ong, Marcus Eng Hock Chakraborty, Bibhas Goldstein, Benjamin Alan Ting, Daniel Shu Wei Vaughan, Roger Liu, Nan |
author_facet | Ning, Yilin Ong, Marcus Eng Hock Chakraborty, Bibhas Goldstein, Benjamin Alan Ting, Daniel Shu Wei Vaughan, Roger Liu, Nan |
author_sort | Ning, Yilin |
collection | PubMed |
description | Interpretable machine learning has been focusing on explaining final models that optimize performance. The state-of-the-art Shapley additive explanations (SHAP) locally explains the variable impact on individual predictions and has recently been extended to provide global assessments across the dataset. Our work further extends “global” assessments to a set of models that are “good enough” and are practically as relevant as the final model to a prediction task. The resulting Shapley variable importance cloud consists of Shapley-based importance measures from each good model and pools information across models to provide an overall importance measure, with uncertainty explicitly quantified to support formal statistical inference. We developed visualizations to highlight the uncertainty and to illustrate its implications to practical inference. Building on a common theoretical basis, our method seamlessly complements the widely adopted SHAP assessments of a single final model to avoid biased inference, which we demonstrate in two experiments using recidivism prediction data and clinical data. |
format | Online Article Text |
id | pubmed-9023900 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Elsevier |
record_format | MEDLINE/PubMed |
spelling | pubmed-90239002022-04-23 Shapley variable importance cloud for interpretable machine learning Ning, Yilin Ong, Marcus Eng Hock Chakraborty, Bibhas Goldstein, Benjamin Alan Ting, Daniel Shu Wei Vaughan, Roger Liu, Nan Patterns (N Y) Article Interpretable machine learning has been focusing on explaining final models that optimize performance. The state-of-the-art Shapley additive explanations (SHAP) locally explains the variable impact on individual predictions and has recently been extended to provide global assessments across the dataset. Our work further extends “global” assessments to a set of models that are “good enough” and are practically as relevant as the final model to a prediction task. The resulting Shapley variable importance cloud consists of Shapley-based importance measures from each good model and pools information across models to provide an overall importance measure, with uncertainty explicitly quantified to support formal statistical inference. We developed visualizations to highlight the uncertainty and to illustrate its implications to practical inference. Building on a common theoretical basis, our method seamlessly complements the widely adopted SHAP assessments of a single final model to avoid biased inference, which we demonstrate in two experiments using recidivism prediction data and clinical data. Elsevier 2022-02-22 /pmc/articles/PMC9023900/ /pubmed/35465224 http://dx.doi.org/10.1016/j.patter.2022.100452 Text en © 2022 The Author(s) https://creativecommons.org/licenses/by/4.0/This is an open access article under the CC BY license (http://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Ning, Yilin Ong, Marcus Eng Hock Chakraborty, Bibhas Goldstein, Benjamin Alan Ting, Daniel Shu Wei Vaughan, Roger Liu, Nan Shapley variable importance cloud for interpretable machine learning |
title | Shapley variable importance cloud for interpretable machine learning |
title_full | Shapley variable importance cloud for interpretable machine learning |
title_fullStr | Shapley variable importance cloud for interpretable machine learning |
title_full_unstemmed | Shapley variable importance cloud for interpretable machine learning |
title_short | Shapley variable importance cloud for interpretable machine learning |
title_sort | shapley variable importance cloud for interpretable machine learning |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9023900/ https://www.ncbi.nlm.nih.gov/pubmed/35465224 http://dx.doi.org/10.1016/j.patter.2022.100452 |
work_keys_str_mv | AT ningyilin shapleyvariableimportancecloudforinterpretablemachinelearning AT ongmarcusenghock shapleyvariableimportancecloudforinterpretablemachinelearning AT chakrabortybibhas shapleyvariableimportancecloudforinterpretablemachinelearning AT goldsteinbenjaminalan shapleyvariableimportancecloudforinterpretablemachinelearning AT tingdanielshuwei shapleyvariableimportancecloudforinterpretablemachinelearning AT vaughanroger shapleyvariableimportancecloudforinterpretablemachinelearning AT liunan shapleyvariableimportancecloudforinterpretablemachinelearning |