Cargando…
Framework for Classifying Explainable Artificial Intelligence (XAI) Algorithms in Clinical Medicine
Artificial intelligence (AI) applied to medicine offers immense promise, in addition to safety and regulatory concerns. Traditional AI produces a core algorithm result, typically without a measure of statistical confidence or an explanation of its biological-theoretical basis. Efforts are underway t...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
JMIR Publications
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10689048/ https://www.ncbi.nlm.nih.gov/pubmed/38046562 http://dx.doi.org/10.2196/50934 |
_version_ | 1785152288440975360 |
---|---|
author | Gniadek, Thomas Kang, Jason Theparee, Talent Krive, Jacob |
author_facet | Gniadek, Thomas Kang, Jason Theparee, Talent Krive, Jacob |
author_sort | Gniadek, Thomas |
collection | PubMed |
description | Artificial intelligence (AI) applied to medicine offers immense promise, in addition to safety and regulatory concerns. Traditional AI produces a core algorithm result, typically without a measure of statistical confidence or an explanation of its biological-theoretical basis. Efforts are underway to develop explainable AI (XAI) algorithms that not only produce a result but also an explanation to support that result. Here we present a framework for classifying XAI algorithms applied to clinical medicine: An algorithm’s clinical scope is defined by whether the core algorithm output leads to observations (eg, tests, imaging, clinical evaluation), interventions (eg, procedures, medications), diagnoses, and prognostication. Explanations are classified by whether they provide empiric statistical information, association with a historical population or populations, or association with an established disease mechanism or mechanisms. XAI implementations can be classified based on whether algorithm training and validation took into account the actions of health care providers in response to the insights and explanations provided or whether training was performed using only the core algorithm output as the end point. Finally, communication modalities used to convey an XAI explanation can be used to classify algorithms and may affect clinical outcomes. This framework can be used when designing, evaluating, and comparing XAI algorithms applied to medicine. |
format | Online Article Text |
id | pubmed-10689048 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | JMIR Publications |
record_format | MEDLINE/PubMed |
spelling | pubmed-106890482023-12-01 Framework for Classifying Explainable Artificial Intelligence (XAI) Algorithms in Clinical Medicine Gniadek, Thomas Kang, Jason Theparee, Talent Krive, Jacob Online J Public Health Inform Viewpoint Artificial intelligence (AI) applied to medicine offers immense promise, in addition to safety and regulatory concerns. Traditional AI produces a core algorithm result, typically without a measure of statistical confidence or an explanation of its biological-theoretical basis. Efforts are underway to develop explainable AI (XAI) algorithms that not only produce a result but also an explanation to support that result. Here we present a framework for classifying XAI algorithms applied to clinical medicine: An algorithm’s clinical scope is defined by whether the core algorithm output leads to observations (eg, tests, imaging, clinical evaluation), interventions (eg, procedures, medications), diagnoses, and prognostication. Explanations are classified by whether they provide empiric statistical information, association with a historical population or populations, or association with an established disease mechanism or mechanisms. XAI implementations can be classified based on whether algorithm training and validation took into account the actions of health care providers in response to the insights and explanations provided or whether training was performed using only the core algorithm output as the end point. Finally, communication modalities used to convey an XAI explanation can be used to classify algorithms and may affect clinical outcomes. This framework can be used when designing, evaluating, and comparing XAI algorithms applied to medicine. JMIR Publications 2023-09-01 /pmc/articles/PMC10689048/ /pubmed/38046562 http://dx.doi.org/10.2196/50934 Text en ©Thomas Gniadek, Jason Kang, Talent Theparee, Jacob Krive. Originally published in the Online Journal of Public Health Informatics (https://ojphi.jmir.org/), 01.09.2023. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Online Journal of Public Health Informatics, is properly cited. The complete bibliographic information, a link to the original publication on https://ojphi.jmir.org/, as well as this copyright and license information must be included. |
spellingShingle | Viewpoint Gniadek, Thomas Kang, Jason Theparee, Talent Krive, Jacob Framework for Classifying Explainable Artificial Intelligence (XAI) Algorithms in Clinical Medicine |
title | Framework for Classifying Explainable Artificial Intelligence (XAI) Algorithms in Clinical Medicine |
title_full | Framework for Classifying Explainable Artificial Intelligence (XAI) Algorithms in Clinical Medicine |
title_fullStr | Framework for Classifying Explainable Artificial Intelligence (XAI) Algorithms in Clinical Medicine |
title_full_unstemmed | Framework for Classifying Explainable Artificial Intelligence (XAI) Algorithms in Clinical Medicine |
title_short | Framework for Classifying Explainable Artificial Intelligence (XAI) Algorithms in Clinical Medicine |
title_sort | framework for classifying explainable artificial intelligence (xai) algorithms in clinical medicine |
topic | Viewpoint |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10689048/ https://www.ncbi.nlm.nih.gov/pubmed/38046562 http://dx.doi.org/10.2196/50934 |
work_keys_str_mv | AT gniadekthomas frameworkforclassifyingexplainableartificialintelligencexaialgorithmsinclinicalmedicine AT kangjason frameworkforclassifyingexplainableartificialintelligencexaialgorithmsinclinicalmedicine AT thepareetalent frameworkforclassifyingexplainableartificialintelligencexaialgorithmsinclinicalmedicine AT krivejacob frameworkforclassifyingexplainableartificialintelligencexaialgorithmsinclinicalmedicine |