Cargando…
Solving the explainable AI conundrum by bridging clinicians’ needs and developers’ goals
Explainable artificial intelligence (XAI) has emerged as a promising solution for addressing the implementation challenges of AI/ML in healthcare. However, little is known about how developers and clinicians interpret XAI and what conflicting goals and requirements they may have. This paper presents...
Autores principales: | , , , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Nature Publishing Group UK
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10202353/ https://www.ncbi.nlm.nih.gov/pubmed/37217779 http://dx.doi.org/10.1038/s41746-023-00837-4 |
_version_ | 1785045425145774080 |
---|---|
author | Bienefeld, Nadine Boss, Jens Michael Lüthy, Rahel Brodbeck, Dominique Azzati, Jan Blaser, Mirco Willms, Jan Keller, Emanuela |
author_facet | Bienefeld, Nadine Boss, Jens Michael Lüthy, Rahel Brodbeck, Dominique Azzati, Jan Blaser, Mirco Willms, Jan Keller, Emanuela |
author_sort | Bienefeld, Nadine |
collection | PubMed |
description | Explainable artificial intelligence (XAI) has emerged as a promising solution for addressing the implementation challenges of AI/ML in healthcare. However, little is known about how developers and clinicians interpret XAI and what conflicting goals and requirements they may have. This paper presents the findings of a longitudinal multi-method study involving 112 developers and clinicians co-designing an XAI solution for a clinical decision support system. Our study identifies three key differences between developer and clinician mental models of XAI, including opposing goals (model interpretability vs. clinical plausibility), different sources of truth (data vs. patient), and the role of exploring new vs. exploiting old knowledge. Based on our findings, we propose design solutions that can help address the XAI conundrum in healthcare, including the use of causal inference models, personalized explanations, and ambidexterity between exploration and exploitation mindsets. Our study highlights the importance of considering the perspectives of both developers and clinicians in the design of XAI systems and provides practical recommendations for improving the effectiveness and usability of XAI in healthcare. |
format | Online Article Text |
id | pubmed-10202353 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Nature Publishing Group UK |
record_format | MEDLINE/PubMed |
spelling | pubmed-102023532023-05-23 Solving the explainable AI conundrum by bridging clinicians’ needs and developers’ goals Bienefeld, Nadine Boss, Jens Michael Lüthy, Rahel Brodbeck, Dominique Azzati, Jan Blaser, Mirco Willms, Jan Keller, Emanuela NPJ Digit Med Brief Communication Explainable artificial intelligence (XAI) has emerged as a promising solution for addressing the implementation challenges of AI/ML in healthcare. However, little is known about how developers and clinicians interpret XAI and what conflicting goals and requirements they may have. This paper presents the findings of a longitudinal multi-method study involving 112 developers and clinicians co-designing an XAI solution for a clinical decision support system. Our study identifies three key differences between developer and clinician mental models of XAI, including opposing goals (model interpretability vs. clinical plausibility), different sources of truth (data vs. patient), and the role of exploring new vs. exploiting old knowledge. Based on our findings, we propose design solutions that can help address the XAI conundrum in healthcare, including the use of causal inference models, personalized explanations, and ambidexterity between exploration and exploitation mindsets. Our study highlights the importance of considering the perspectives of both developers and clinicians in the design of XAI systems and provides practical recommendations for improving the effectiveness and usability of XAI in healthcare. Nature Publishing Group UK 2023-05-22 /pmc/articles/PMC10202353/ /pubmed/37217779 http://dx.doi.org/10.1038/s41746-023-00837-4 Text en © The Author(s) 2023 https://creativecommons.org/licenses/by/4.0/Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) . |
spellingShingle | Brief Communication Bienefeld, Nadine Boss, Jens Michael Lüthy, Rahel Brodbeck, Dominique Azzati, Jan Blaser, Mirco Willms, Jan Keller, Emanuela Solving the explainable AI conundrum by bridging clinicians’ needs and developers’ goals |
title | Solving the explainable AI conundrum by bridging clinicians’ needs and developers’ goals |
title_full | Solving the explainable AI conundrum by bridging clinicians’ needs and developers’ goals |
title_fullStr | Solving the explainable AI conundrum by bridging clinicians’ needs and developers’ goals |
title_full_unstemmed | Solving the explainable AI conundrum by bridging clinicians’ needs and developers’ goals |
title_short | Solving the explainable AI conundrum by bridging clinicians’ needs and developers’ goals |
title_sort | solving the explainable ai conundrum by bridging clinicians’ needs and developers’ goals |
topic | Brief Communication |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10202353/ https://www.ncbi.nlm.nih.gov/pubmed/37217779 http://dx.doi.org/10.1038/s41746-023-00837-4 |
work_keys_str_mv | AT bienefeldnadine solvingtheexplainableaiconundrumbybridgingcliniciansneedsanddevelopersgoals AT bossjensmichael solvingtheexplainableaiconundrumbybridgingcliniciansneedsanddevelopersgoals AT luthyrahel solvingtheexplainableaiconundrumbybridgingcliniciansneedsanddevelopersgoals AT brodbeckdominique solvingtheexplainableaiconundrumbybridgingcliniciansneedsanddevelopersgoals AT azzatijan solvingtheexplainableaiconundrumbybridgingcliniciansneedsanddevelopersgoals AT blasermirco solvingtheexplainableaiconundrumbybridgingcliniciansneedsanddevelopersgoals AT willmsjan solvingtheexplainableaiconundrumbybridgingcliniciansneedsanddevelopersgoals AT kelleremanuela solvingtheexplainableaiconundrumbybridgingcliniciansneedsanddevelopersgoals |