Cargando…

Counterfactual Explanation of Brain Activity Classifiers Using Image-To-Image Transfer by Generative Adversarial Network

Deep neural networks (DNNs) can accurately decode task-related information from brain activations. However, because of the non-linearity of DNNs, it is generally difficult to explain how and why they assign certain behavioral tasks to given brain activations, either correctly or incorrectly. One of...

Descripción completa

Detalles Bibliográficos
Autores principales: Matsui, Teppei, Taki, Masato, Pham, Trung Quang, Chikazoe, Junichi, Jimura, Koji
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8966478/
https://www.ncbi.nlm.nih.gov/pubmed/35369003
http://dx.doi.org/10.3389/fninf.2021.802938
_version_ 1784678654770413568
author Matsui, Teppei
Taki, Masato
Pham, Trung Quang
Chikazoe, Junichi
Jimura, Koji
author_facet Matsui, Teppei
Taki, Masato
Pham, Trung Quang
Chikazoe, Junichi
Jimura, Koji
author_sort Matsui, Teppei
collection PubMed
description Deep neural networks (DNNs) can accurately decode task-related information from brain activations. However, because of the non-linearity of DNNs, it is generally difficult to explain how and why they assign certain behavioral tasks to given brain activations, either correctly or incorrectly. One of the promising approaches for explaining such a black-box system is counterfactual explanation. In this framework, the behavior of a black-box system is explained by comparing real data and realistic synthetic data that are specifically generated such that the black-box system outputs an unreal outcome. The explanation of the system's decision can be explained by directly comparing the real and synthetic data. Recently, by taking advantage of advances in DNN-based image-to-image translation, several studies successfully applied counterfactual explanation to image domains. In principle, the same approach could be used in functional magnetic resonance imaging (fMRI) data. Because fMRI datasets often contain multiple classes (e.g., multiple behavioral tasks), the image-to-image transformation applicable to counterfactual explanation needs to learn mapping among multiple classes simultaneously. Recently, a new generative neural network (StarGAN) that enables image-to-image transformation among multiple classes has been developed. By adapting StarGAN with some modifications, here, we introduce a novel generative DNN (counterfactual activation generator, CAG) that can provide counterfactual explanations for DNN-based classifiers of brain activations. Importantly, CAG can simultaneously handle image transformation among all the seven classes in a publicly available fMRI dataset. Thus, CAG could provide a counterfactual explanation of DNN-based multiclass classifiers of brain activations. Furthermore, iterative applications of CAG were able to enhance and extract subtle spatial brain activity patterns that affected the classifier's decisions. Together, these results demonstrate that the counterfactual explanation based on image-to-image transformation would be a promising approach to understand and extend the current application of DNNs in fMRI analyses.
format Online
Article
Text
id pubmed-8966478
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-89664782022-03-31 Counterfactual Explanation of Brain Activity Classifiers Using Image-To-Image Transfer by Generative Adversarial Network Matsui, Teppei Taki, Masato Pham, Trung Quang Chikazoe, Junichi Jimura, Koji Front Neuroinform Neuroscience Deep neural networks (DNNs) can accurately decode task-related information from brain activations. However, because of the non-linearity of DNNs, it is generally difficult to explain how and why they assign certain behavioral tasks to given brain activations, either correctly or incorrectly. One of the promising approaches for explaining such a black-box system is counterfactual explanation. In this framework, the behavior of a black-box system is explained by comparing real data and realistic synthetic data that are specifically generated such that the black-box system outputs an unreal outcome. The explanation of the system's decision can be explained by directly comparing the real and synthetic data. Recently, by taking advantage of advances in DNN-based image-to-image translation, several studies successfully applied counterfactual explanation to image domains. In principle, the same approach could be used in functional magnetic resonance imaging (fMRI) data. Because fMRI datasets often contain multiple classes (e.g., multiple behavioral tasks), the image-to-image transformation applicable to counterfactual explanation needs to learn mapping among multiple classes simultaneously. Recently, a new generative neural network (StarGAN) that enables image-to-image transformation among multiple classes has been developed. By adapting StarGAN with some modifications, here, we introduce a novel generative DNN (counterfactual activation generator, CAG) that can provide counterfactual explanations for DNN-based classifiers of brain activations. Importantly, CAG can simultaneously handle image transformation among all the seven classes in a publicly available fMRI dataset. Thus, CAG could provide a counterfactual explanation of DNN-based multiclass classifiers of brain activations. Furthermore, iterative applications of CAG were able to enhance and extract subtle spatial brain activity patterns that affected the classifier's decisions. Together, these results demonstrate that the counterfactual explanation based on image-to-image transformation would be a promising approach to understand and extend the current application of DNNs in fMRI analyses. Frontiers Media S.A. 2022-03-16 /pmc/articles/PMC8966478/ /pubmed/35369003 http://dx.doi.org/10.3389/fninf.2021.802938 Text en Copyright © 2022 Matsui, Taki, Pham, Chikazoe and Jimura. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Matsui, Teppei
Taki, Masato
Pham, Trung Quang
Chikazoe, Junichi
Jimura, Koji
Counterfactual Explanation of Brain Activity Classifiers Using Image-To-Image Transfer by Generative Adversarial Network
title Counterfactual Explanation of Brain Activity Classifiers Using Image-To-Image Transfer by Generative Adversarial Network
title_full Counterfactual Explanation of Brain Activity Classifiers Using Image-To-Image Transfer by Generative Adversarial Network
title_fullStr Counterfactual Explanation of Brain Activity Classifiers Using Image-To-Image Transfer by Generative Adversarial Network
title_full_unstemmed Counterfactual Explanation of Brain Activity Classifiers Using Image-To-Image Transfer by Generative Adversarial Network
title_short Counterfactual Explanation of Brain Activity Classifiers Using Image-To-Image Transfer by Generative Adversarial Network
title_sort counterfactual explanation of brain activity classifiers using image-to-image transfer by generative adversarial network
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8966478/
https://www.ncbi.nlm.nih.gov/pubmed/35369003
http://dx.doi.org/10.3389/fninf.2021.802938
work_keys_str_mv AT matsuiteppei counterfactualexplanationofbrainactivityclassifiersusingimagetoimagetransferbygenerativeadversarialnetwork
AT takimasato counterfactualexplanationofbrainactivityclassifiersusingimagetoimagetransferbygenerativeadversarialnetwork
AT phamtrungquang counterfactualexplanationofbrainactivityclassifiersusingimagetoimagetransferbygenerativeadversarialnetwork
AT chikazoejunichi counterfactualexplanationofbrainactivityclassifiersusingimagetoimagetransferbygenerativeadversarialnetwork
AT jimurakoji counterfactualexplanationofbrainactivityclassifiersusingimagetoimagetransferbygenerativeadversarialnetwork