Cargando…
Interpretable machine learning with tree-based shapley additive explanations: Application to metabolomics datasets for binary classification
Machine learning (ML) models are used in clinical metabolomics studies most notably for biomarker discoveries, to identify metabolites that discriminate between a case and control group. To improve understanding of the underlying biomedical problem and to bolster confidence in these discoveries, mod...
Autor principal: | |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Public Library of Science
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10159207/ https://www.ncbi.nlm.nih.gov/pubmed/37141218 http://dx.doi.org/10.1371/journal.pone.0284315 |
_version_ | 1785037086561140736 |
---|---|
author | Bifarin, Olatomiwa O. |
author_facet | Bifarin, Olatomiwa O. |
author_sort | Bifarin, Olatomiwa O. |
collection | PubMed |
description | Machine learning (ML) models are used in clinical metabolomics studies most notably for biomarker discoveries, to identify metabolites that discriminate between a case and control group. To improve understanding of the underlying biomedical problem and to bolster confidence in these discoveries, model interpretability is germane. In metabolomics, partial least square discriminant analysis (PLS-DA) and its variants are widely used, partly due to the model’s interpretability with the Variable Influence in Projection (VIP) scores, a global interpretable method. Herein, Tree-based Shapley Additive explanations (SHAP), an interpretable ML method grounded in game theory, was used to explain ML models with local explanation properties. In this study, ML experiments (binary classification) were conducted for three published metabolomics datasets using PLS-DA, random forests, gradient boosting, and extreme gradient boosting (XGBoost). Using one of the datasets, PLS-DA model was explained using VIP scores, while one of the best-performing models, a random forest model, was interpreted using Tree SHAP. The results show that SHAP has a more explanation depth than PLS-DA’s VIP, making it a powerful method for rationalizing machine learning predictions from metabolomics studies. |
format | Online Article Text |
id | pubmed-10159207 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Public Library of Science |
record_format | MEDLINE/PubMed |
spelling | pubmed-101592072023-05-05 Interpretable machine learning with tree-based shapley additive explanations: Application to metabolomics datasets for binary classification Bifarin, Olatomiwa O. PLoS One Research Article Machine learning (ML) models are used in clinical metabolomics studies most notably for biomarker discoveries, to identify metabolites that discriminate between a case and control group. To improve understanding of the underlying biomedical problem and to bolster confidence in these discoveries, model interpretability is germane. In metabolomics, partial least square discriminant analysis (PLS-DA) and its variants are widely used, partly due to the model’s interpretability with the Variable Influence in Projection (VIP) scores, a global interpretable method. Herein, Tree-based Shapley Additive explanations (SHAP), an interpretable ML method grounded in game theory, was used to explain ML models with local explanation properties. In this study, ML experiments (binary classification) were conducted for three published metabolomics datasets using PLS-DA, random forests, gradient boosting, and extreme gradient boosting (XGBoost). Using one of the datasets, PLS-DA model was explained using VIP scores, while one of the best-performing models, a random forest model, was interpreted using Tree SHAP. The results show that SHAP has a more explanation depth than PLS-DA’s VIP, making it a powerful method for rationalizing machine learning predictions from metabolomics studies. Public Library of Science 2023-05-04 /pmc/articles/PMC10159207/ /pubmed/37141218 http://dx.doi.org/10.1371/journal.pone.0284315 Text en © 2023 Olatomiwa O. Bifarin https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. |
spellingShingle | Research Article Bifarin, Olatomiwa O. Interpretable machine learning with tree-based shapley additive explanations: Application to metabolomics datasets for binary classification |
title | Interpretable machine learning with tree-based shapley additive explanations: Application to metabolomics datasets for binary classification |
title_full | Interpretable machine learning with tree-based shapley additive explanations: Application to metabolomics datasets for binary classification |
title_fullStr | Interpretable machine learning with tree-based shapley additive explanations: Application to metabolomics datasets for binary classification |
title_full_unstemmed | Interpretable machine learning with tree-based shapley additive explanations: Application to metabolomics datasets for binary classification |
title_short | Interpretable machine learning with tree-based shapley additive explanations: Application to metabolomics datasets for binary classification |
title_sort | interpretable machine learning with tree-based shapley additive explanations: application to metabolomics datasets for binary classification |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10159207/ https://www.ncbi.nlm.nih.gov/pubmed/37141218 http://dx.doi.org/10.1371/journal.pone.0284315 |
work_keys_str_mv | AT bifarinolatomiwao interpretablemachinelearningwithtreebasedshapleyadditiveexplanationsapplicationtometabolomicsdatasetsforbinaryclassification |