Cargando…

Interpretable brain disease classification and relevance-guided deep learning

Deep neural networks are increasingly used for neurological disease classification by MRI, but the networks’ decisions are not easily interpretable by humans. Heat mapping by deep Taylor decomposition revealed that (potentially misleading) image features even outside of the brain tissue are crucial...

Descripción completa

Detalles Bibliográficos
Autores principales: Tinauer, Christian, Heber, Stefan, Pirpamer, Lukas, Damulina, Anna, Schmidt, Reinhold, Stollberger, Rudolf, Ropele, Stefan, Langkammer, Christian
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Nature Publishing Group UK 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9691637/
https://www.ncbi.nlm.nih.gov/pubmed/36424437
http://dx.doi.org/10.1038/s41598-022-24541-7
_version_ 1784837068019466240
author Tinauer, Christian
Heber, Stefan
Pirpamer, Lukas
Damulina, Anna
Schmidt, Reinhold
Stollberger, Rudolf
Ropele, Stefan
Langkammer, Christian
author_facet Tinauer, Christian
Heber, Stefan
Pirpamer, Lukas
Damulina, Anna
Schmidt, Reinhold
Stollberger, Rudolf
Ropele, Stefan
Langkammer, Christian
author_sort Tinauer, Christian
collection PubMed
description Deep neural networks are increasingly used for neurological disease classification by MRI, but the networks’ decisions are not easily interpretable by humans. Heat mapping by deep Taylor decomposition revealed that (potentially misleading) image features even outside of the brain tissue are crucial for the classifier’s decision. We propose a regularization technique to train convolutional neural network (CNN) classifiers utilizing relevance-guided heat maps calculated online during training. The method was applied using T1-weighted MR images from 128 subjects with Alzheimer’s disease (mean age = 71.9 ± 8.5 years) and 290 control subjects (mean age = 71.3 ± 6.4 years). The developed relevance-guided framework achieves higher classification accuracies than conventional CNNs but more importantly, it relies on less but more relevant and physiological plausible voxels within brain tissue. Additionally, preprocessing effects from skull stripping and registration are mitigated. With the interpretability of the decision mechanisms underlying CNNs, these results challenge the notion that unprocessed T1-weighted brain MR images in standard CNNs yield higher classification accuracy in Alzheimer’s disease than solely atrophy.
format Online
Article
Text
id pubmed-9691637
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Nature Publishing Group UK
record_format MEDLINE/PubMed
spelling pubmed-96916372022-11-26 Interpretable brain disease classification and relevance-guided deep learning Tinauer, Christian Heber, Stefan Pirpamer, Lukas Damulina, Anna Schmidt, Reinhold Stollberger, Rudolf Ropele, Stefan Langkammer, Christian Sci Rep Article Deep neural networks are increasingly used for neurological disease classification by MRI, but the networks’ decisions are not easily interpretable by humans. Heat mapping by deep Taylor decomposition revealed that (potentially misleading) image features even outside of the brain tissue are crucial for the classifier’s decision. We propose a regularization technique to train convolutional neural network (CNN) classifiers utilizing relevance-guided heat maps calculated online during training. The method was applied using T1-weighted MR images from 128 subjects with Alzheimer’s disease (mean age = 71.9 ± 8.5 years) and 290 control subjects (mean age = 71.3 ± 6.4 years). The developed relevance-guided framework achieves higher classification accuracies than conventional CNNs but more importantly, it relies on less but more relevant and physiological plausible voxels within brain tissue. Additionally, preprocessing effects from skull stripping and registration are mitigated. With the interpretability of the decision mechanisms underlying CNNs, these results challenge the notion that unprocessed T1-weighted brain MR images in standard CNNs yield higher classification accuracy in Alzheimer’s disease than solely atrophy. Nature Publishing Group UK 2022-11-24 /pmc/articles/PMC9691637/ /pubmed/36424437 http://dx.doi.org/10.1038/s41598-022-24541-7 Text en © The Author(s) 2022 https://creativecommons.org/licenses/by/4.0/Open AccessThis article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) .
spellingShingle Article
Tinauer, Christian
Heber, Stefan
Pirpamer, Lukas
Damulina, Anna
Schmidt, Reinhold
Stollberger, Rudolf
Ropele, Stefan
Langkammer, Christian
Interpretable brain disease classification and relevance-guided deep learning
title Interpretable brain disease classification and relevance-guided deep learning
title_full Interpretable brain disease classification and relevance-guided deep learning
title_fullStr Interpretable brain disease classification and relevance-guided deep learning
title_full_unstemmed Interpretable brain disease classification and relevance-guided deep learning
title_short Interpretable brain disease classification and relevance-guided deep learning
title_sort interpretable brain disease classification and relevance-guided deep learning
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9691637/
https://www.ncbi.nlm.nih.gov/pubmed/36424437
http://dx.doi.org/10.1038/s41598-022-24541-7
work_keys_str_mv AT tinauerchristian interpretablebraindiseaseclassificationandrelevanceguideddeeplearning
AT heberstefan interpretablebraindiseaseclassificationandrelevanceguideddeeplearning
AT pirpamerlukas interpretablebraindiseaseclassificationandrelevanceguideddeeplearning
AT damulinaanna interpretablebraindiseaseclassificationandrelevanceguideddeeplearning
AT schmidtreinhold interpretablebraindiseaseclassificationandrelevanceguideddeeplearning
AT stollbergerrudolf interpretablebraindiseaseclassificationandrelevanceguideddeeplearning
AT ropelestefan interpretablebraindiseaseclassificationandrelevanceguideddeeplearning
AT langkammerchristian interpretablebraindiseaseclassificationandrelevanceguideddeeplearning