Cargando…
Dissection of medical AI reasoning processes via physician and generative-AI collaboration
Despite the proliferation and clinical deployment of artificial intelligence (AI)-based medical software devices, most remain black boxes that are uninterpretable to key stakeholders including patients, physicians, and even the developers of the devices. Here, we present a general model auditing fra...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Cold Spring Harbor Laboratory
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10246034/ https://www.ncbi.nlm.nih.gov/pubmed/37292705 http://dx.doi.org/10.1101/2023.05.12.23289878 |
_version_ | 1785054965550546944 |
---|---|
author | DeGrave, Alex J. Cai, Zhuo Ran Janizek, Joseph D. Daneshjou, Roxana Lee, Su-In |
author_facet | DeGrave, Alex J. Cai, Zhuo Ran Janizek, Joseph D. Daneshjou, Roxana Lee, Su-In |
author_sort | DeGrave, Alex J. |
collection | PubMed |
description | Despite the proliferation and clinical deployment of artificial intelligence (AI)-based medical software devices, most remain black boxes that are uninterpretable to key stakeholders including patients, physicians, and even the developers of the devices. Here, we present a general model auditing framework that combines insights from medical experts with a highly expressive form of explainable AI that leverages generative models, to understand the reasoning processes of AI devices. We then apply this framework to generate the first thorough, medically interpretable picture of the reasoning processes of machine-learning–based medical image AI. In our synergistic framework, a generative model first renders “counterfactual” medical images, which in essence visually represent the reasoning process of a medical AI device, and then physicians translate these counterfactual images to medically meaningful features. As our use case, we audit five high-profile AI devices in dermatology, an area of particular interest since dermatology AI devices are beginning to achieve deployment globally. We reveal how dermatology AI devices rely both on features used by human dermatologists, such as lesional pigmentation patterns, as well as multiple, previously unreported, potentially undesirable features, such as background skin texture and image color balance. Our study also sets a precedent for the rigorous application of explainable AI to understand AI in any specialized domain and provides a means for practitioners, clinicians, and regulators to uncloak AI’s powerful but previously enigmatic reasoning processes in a medically understandable way. |
format | Online Article Text |
id | pubmed-10246034 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Cold Spring Harbor Laboratory |
record_format | MEDLINE/PubMed |
spelling | pubmed-102460342023-06-08 Dissection of medical AI reasoning processes via physician and generative-AI collaboration DeGrave, Alex J. Cai, Zhuo Ran Janizek, Joseph D. Daneshjou, Roxana Lee, Su-In medRxiv Article Despite the proliferation and clinical deployment of artificial intelligence (AI)-based medical software devices, most remain black boxes that are uninterpretable to key stakeholders including patients, physicians, and even the developers of the devices. Here, we present a general model auditing framework that combines insights from medical experts with a highly expressive form of explainable AI that leverages generative models, to understand the reasoning processes of AI devices. We then apply this framework to generate the first thorough, medically interpretable picture of the reasoning processes of machine-learning–based medical image AI. In our synergistic framework, a generative model first renders “counterfactual” medical images, which in essence visually represent the reasoning process of a medical AI device, and then physicians translate these counterfactual images to medically meaningful features. As our use case, we audit five high-profile AI devices in dermatology, an area of particular interest since dermatology AI devices are beginning to achieve deployment globally. We reveal how dermatology AI devices rely both on features used by human dermatologists, such as lesional pigmentation patterns, as well as multiple, previously unreported, potentially undesirable features, such as background skin texture and image color balance. Our study also sets a precedent for the rigorous application of explainable AI to understand AI in any specialized domain and provides a means for practitioners, clinicians, and regulators to uncloak AI’s powerful but previously enigmatic reasoning processes in a medically understandable way. Cold Spring Harbor Laboratory 2023-05-16 /pmc/articles/PMC10246034/ /pubmed/37292705 http://dx.doi.org/10.1101/2023.05.12.23289878 Text en https://creativecommons.org/licenses/by-nc/4.0/This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License (https://creativecommons.org/licenses/by-nc/4.0/) , which allows reusers to distribute, remix, adapt, and build upon the material in any medium or format for noncommercial purposes only, and only so long as attribution is given to the creator. |
spellingShingle | Article DeGrave, Alex J. Cai, Zhuo Ran Janizek, Joseph D. Daneshjou, Roxana Lee, Su-In Dissection of medical AI reasoning processes via physician and generative-AI collaboration |
title | Dissection of medical AI reasoning processes via physician and generative-AI collaboration |
title_full | Dissection of medical AI reasoning processes via physician and generative-AI collaboration |
title_fullStr | Dissection of medical AI reasoning processes via physician and generative-AI collaboration |
title_full_unstemmed | Dissection of medical AI reasoning processes via physician and generative-AI collaboration |
title_short | Dissection of medical AI reasoning processes via physician and generative-AI collaboration |
title_sort | dissection of medical ai reasoning processes via physician and generative-ai collaboration |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10246034/ https://www.ncbi.nlm.nih.gov/pubmed/37292705 http://dx.doi.org/10.1101/2023.05.12.23289878 |
work_keys_str_mv | AT degravealexj dissectionofmedicalaireasoningprocessesviaphysicianandgenerativeaicollaboration AT caizhuoran dissectionofmedicalaireasoningprocessesviaphysicianandgenerativeaicollaboration AT janizekjosephd dissectionofmedicalaireasoningprocessesviaphysicianandgenerativeaicollaboration AT daneshjouroxana dissectionofmedicalaireasoningprocessesviaphysicianandgenerativeaicollaboration AT leesuin dissectionofmedicalaireasoningprocessesviaphysicianandgenerativeaicollaboration |