Cargando…

Utilization of model-agnostic explainable artificial intelligence frameworks in oncology: a narrative review

BACKGROUND AND OBJECTIVE: Machine learning (ML) models are increasingly being utilized in oncology research for use in the clinic. However, while more complicated models may provide improvements in predictive or prognostic power, a hurdle to their adoption are limits of model interpretability, where...

Descripción completa

Detalles Bibliográficos
Autores principales: Ladbury, Colton, Zarinshenas, Reza, Semwal, Hemal, Tam, Andrew, Vaidehi, Nagarajan, Rodin, Andrei S., Liu, An, Glaser, Scott, Salgia, Ravi, Amini, Arya
Formato: Online Artículo Texto
Lenguaje:English
Publicado: AME Publishing Company 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9641128/
https://www.ncbi.nlm.nih.gov/pubmed/36388027
http://dx.doi.org/10.21037/tcr-22-1626
_version_ 1784826025571516416
author Ladbury, Colton
Zarinshenas, Reza
Semwal, Hemal
Tam, Andrew
Vaidehi, Nagarajan
Rodin, Andrei S.
Liu, An
Glaser, Scott
Salgia, Ravi
Amini, Arya
author_facet Ladbury, Colton
Zarinshenas, Reza
Semwal, Hemal
Tam, Andrew
Vaidehi, Nagarajan
Rodin, Andrei S.
Liu, An
Glaser, Scott
Salgia, Ravi
Amini, Arya
author_sort Ladbury, Colton
collection PubMed
description BACKGROUND AND OBJECTIVE: Machine learning (ML) models are increasingly being utilized in oncology research for use in the clinic. However, while more complicated models may provide improvements in predictive or prognostic power, a hurdle to their adoption are limits of model interpretability, wherein the inner workings can be perceived as a “black box”. Explainable artificial intelligence (XAI) frameworks including Local Interpretable Model-agnostic Explanations (LIME) and SHapley Additive exPlanations (SHAP) are novel, model-agnostic approaches that aim to provide insight into the inner workings of the “black box” by producing quantitative visualizations of how model predictions are calculated. In doing so, XAI can transform complicated ML models into easily understandable charts and interpretable sets of rules, which can give providers with an intuitive understanding of the knowledge generated, thus facilitating the deployment of such models in routine clinical workflows. METHODS: We performed a comprehensive, non-systematic review of the latest literature to define use cases of model-agnostic XAI frameworks in oncologic research. The examined database was PubMed/MEDLINE. The last search was run on May 1, 2022. KEY CONTENT AND FINDINGS: In this review, we identified several fields in oncology research where ML models and XAI were utilized to improve interpretability, including prognostication, diagnosis, radiomics, pathology, treatment selection, radiation treatment workflows, and epidemiology. Within these fields, XAI facilitates determination of feature importance in the overall model, visualization of relationships and/or interactions, evaluation of how individual predictions are produced, feature selection, identification of prognostic and/or predictive thresholds, and overall confidence in the models, among other benefits. These examples provide a basis for future work to expand on, which can facilitate adoption in the clinic when the complexity of such modeling would otherwise be prohibitive. CONCLUSIONS: Model-agnostic XAI frameworks offer an intuitive and effective means of describing oncology ML models, with applications including prognostication and determination of optimal treatment regimens. Using such frameworks presents an opportunity to improve understanding of ML models, which is a critical step to their adoption in the clinic.
format Online
Article
Text
id pubmed-9641128
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher AME Publishing Company
record_format MEDLINE/PubMed
spelling pubmed-96411282022-11-15 Utilization of model-agnostic explainable artificial intelligence frameworks in oncology: a narrative review Ladbury, Colton Zarinshenas, Reza Semwal, Hemal Tam, Andrew Vaidehi, Nagarajan Rodin, Andrei S. Liu, An Glaser, Scott Salgia, Ravi Amini, Arya Transl Cancer Res Review Article BACKGROUND AND OBJECTIVE: Machine learning (ML) models are increasingly being utilized in oncology research for use in the clinic. However, while more complicated models may provide improvements in predictive or prognostic power, a hurdle to their adoption are limits of model interpretability, wherein the inner workings can be perceived as a “black box”. Explainable artificial intelligence (XAI) frameworks including Local Interpretable Model-agnostic Explanations (LIME) and SHapley Additive exPlanations (SHAP) are novel, model-agnostic approaches that aim to provide insight into the inner workings of the “black box” by producing quantitative visualizations of how model predictions are calculated. In doing so, XAI can transform complicated ML models into easily understandable charts and interpretable sets of rules, which can give providers with an intuitive understanding of the knowledge generated, thus facilitating the deployment of such models in routine clinical workflows. METHODS: We performed a comprehensive, non-systematic review of the latest literature to define use cases of model-agnostic XAI frameworks in oncologic research. The examined database was PubMed/MEDLINE. The last search was run on May 1, 2022. KEY CONTENT AND FINDINGS: In this review, we identified several fields in oncology research where ML models and XAI were utilized to improve interpretability, including prognostication, diagnosis, radiomics, pathology, treatment selection, radiation treatment workflows, and epidemiology. Within these fields, XAI facilitates determination of feature importance in the overall model, visualization of relationships and/or interactions, evaluation of how individual predictions are produced, feature selection, identification of prognostic and/or predictive thresholds, and overall confidence in the models, among other benefits. These examples provide a basis for future work to expand on, which can facilitate adoption in the clinic when the complexity of such modeling would otherwise be prohibitive. CONCLUSIONS: Model-agnostic XAI frameworks offer an intuitive and effective means of describing oncology ML models, with applications including prognostication and determination of optimal treatment regimens. Using such frameworks presents an opportunity to improve understanding of ML models, which is a critical step to their adoption in the clinic. AME Publishing Company 2022-10 /pmc/articles/PMC9641128/ /pubmed/36388027 http://dx.doi.org/10.21037/tcr-22-1626 Text en 2022 Translational Cancer Research. All rights reserved. https://creativecommons.org/licenses/by-nc-nd/4.0/Open Access Statement: This is an Open Access article distributed in accordance with the Creative Commons Attribution-NonCommercial-NoDerivs 4.0 International License (CC BY-NC-ND 4.0), which permits the non-commercial replication and distribution of the article with the strict proviso that no changes or edits are made and the original work is properly cited (including links to both the formal publication through the relevant DOI and the license). See: https://creativecommons.org/licenses/by-nc-nd/4.0 (https://creativecommons.org/licenses/by-nc-nd/4.0/) .
spellingShingle Review Article
Ladbury, Colton
Zarinshenas, Reza
Semwal, Hemal
Tam, Andrew
Vaidehi, Nagarajan
Rodin, Andrei S.
Liu, An
Glaser, Scott
Salgia, Ravi
Amini, Arya
Utilization of model-agnostic explainable artificial intelligence frameworks in oncology: a narrative review
title Utilization of model-agnostic explainable artificial intelligence frameworks in oncology: a narrative review
title_full Utilization of model-agnostic explainable artificial intelligence frameworks in oncology: a narrative review
title_fullStr Utilization of model-agnostic explainable artificial intelligence frameworks in oncology: a narrative review
title_full_unstemmed Utilization of model-agnostic explainable artificial intelligence frameworks in oncology: a narrative review
title_short Utilization of model-agnostic explainable artificial intelligence frameworks in oncology: a narrative review
title_sort utilization of model-agnostic explainable artificial intelligence frameworks in oncology: a narrative review
topic Review Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9641128/
https://www.ncbi.nlm.nih.gov/pubmed/36388027
http://dx.doi.org/10.21037/tcr-22-1626
work_keys_str_mv AT ladburycolton utilizationofmodelagnosticexplainableartificialintelligenceframeworksinoncologyanarrativereview
AT zarinshenasreza utilizationofmodelagnosticexplainableartificialintelligenceframeworksinoncologyanarrativereview
AT semwalhemal utilizationofmodelagnosticexplainableartificialintelligenceframeworksinoncologyanarrativereview
AT tamandrew utilizationofmodelagnosticexplainableartificialintelligenceframeworksinoncologyanarrativereview
AT vaidehinagarajan utilizationofmodelagnosticexplainableartificialintelligenceframeworksinoncologyanarrativereview
AT rodinandreis utilizationofmodelagnosticexplainableartificialintelligenceframeworksinoncologyanarrativereview
AT liuan utilizationofmodelagnosticexplainableartificialintelligenceframeworksinoncologyanarrativereview
AT glaserscott utilizationofmodelagnosticexplainableartificialintelligenceframeworksinoncologyanarrativereview
AT salgiaravi utilizationofmodelagnosticexplainableartificialintelligenceframeworksinoncologyanarrativereview
AT aminiarya utilizationofmodelagnosticexplainableartificialintelligenceframeworksinoncologyanarrativereview