Cargando…

Resolving challenges in deep learning-based analyses of histopathological images using explanation methods

Deep learning has recently gained popularity in digital pathology due to its high prediction quality. However, the medical domain requires explanation and insight for a better understanding beyond standard quantitative performance evaluation. Recently, many explanation methods have emerged. This wor...

Descripción completa

Detalles Bibliográficos
Autores principales: Hägele, Miriam, Seegerer, Philipp, Lapuschkin, Sebastian, Bockmayr, Michael, Samek, Wojciech, Klauschen, Frederick, Müller, Klaus-Robert, Binder, Alexander
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Nature Publishing Group UK 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7156509/
https://www.ncbi.nlm.nih.gov/pubmed/32286358
http://dx.doi.org/10.1038/s41598-020-62724-2
_version_ 1783522223173140480
author Hägele, Miriam
Seegerer, Philipp
Lapuschkin, Sebastian
Bockmayr, Michael
Samek, Wojciech
Klauschen, Frederick
Müller, Klaus-Robert
Binder, Alexander
author_facet Hägele, Miriam
Seegerer, Philipp
Lapuschkin, Sebastian
Bockmayr, Michael
Samek, Wojciech
Klauschen, Frederick
Müller, Klaus-Robert
Binder, Alexander
author_sort Hägele, Miriam
collection PubMed
description Deep learning has recently gained popularity in digital pathology due to its high prediction quality. However, the medical domain requires explanation and insight for a better understanding beyond standard quantitative performance evaluation. Recently, many explanation methods have emerged. This work shows how heatmaps generated by these explanation methods allow to resolve common challenges encountered in deep learning-based digital histopathology analyses. We elaborate on biases which are typically inherent in histopathological image data. In the binary classification task of tumour tissue discrimination in publicly available haematoxylin-eosin-stained images of various tumour entities, we investigate three types of biases: (1) biases which affect the entire dataset, (2) biases which are by chance correlated with class labels and (3) sampling biases. While standard analyses focus on patch-level evaluation, we advocate pixel-wise heatmaps, which offer a more precise and versatile diagnostic instrument. This insight is shown to not only be helpful to detect but also to remove the effects of common hidden biases, which improves generalisation within and across datasets. For example, we could see a trend of improved area under the receiver operating characteristic (ROC) curve by 5% when reducing a labelling bias. Explanation techniques are thus demonstrated to be a helpful and highly relevant tool for the development and the deployment phases within the life cycle of real-world applications in digital pathology.
format Online
Article
Text
id pubmed-7156509
institution National Center for Biotechnology Information
language English
publishDate 2020
publisher Nature Publishing Group UK
record_format MEDLINE/PubMed
spelling pubmed-71565092020-04-19 Resolving challenges in deep learning-based analyses of histopathological images using explanation methods Hägele, Miriam Seegerer, Philipp Lapuschkin, Sebastian Bockmayr, Michael Samek, Wojciech Klauschen, Frederick Müller, Klaus-Robert Binder, Alexander Sci Rep Article Deep learning has recently gained popularity in digital pathology due to its high prediction quality. However, the medical domain requires explanation and insight for a better understanding beyond standard quantitative performance evaluation. Recently, many explanation methods have emerged. This work shows how heatmaps generated by these explanation methods allow to resolve common challenges encountered in deep learning-based digital histopathology analyses. We elaborate on biases which are typically inherent in histopathological image data. In the binary classification task of tumour tissue discrimination in publicly available haematoxylin-eosin-stained images of various tumour entities, we investigate three types of biases: (1) biases which affect the entire dataset, (2) biases which are by chance correlated with class labels and (3) sampling biases. While standard analyses focus on patch-level evaluation, we advocate pixel-wise heatmaps, which offer a more precise and versatile diagnostic instrument. This insight is shown to not only be helpful to detect but also to remove the effects of common hidden biases, which improves generalisation within and across datasets. For example, we could see a trend of improved area under the receiver operating characteristic (ROC) curve by 5% when reducing a labelling bias. Explanation techniques are thus demonstrated to be a helpful and highly relevant tool for the development and the deployment phases within the life cycle of real-world applications in digital pathology. Nature Publishing Group UK 2020-04-14 /pmc/articles/PMC7156509/ /pubmed/32286358 http://dx.doi.org/10.1038/s41598-020-62724-2 Text en © The Author(s) 2020 Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.
spellingShingle Article
Hägele, Miriam
Seegerer, Philipp
Lapuschkin, Sebastian
Bockmayr, Michael
Samek, Wojciech
Klauschen, Frederick
Müller, Klaus-Robert
Binder, Alexander
Resolving challenges in deep learning-based analyses of histopathological images using explanation methods
title Resolving challenges in deep learning-based analyses of histopathological images using explanation methods
title_full Resolving challenges in deep learning-based analyses of histopathological images using explanation methods
title_fullStr Resolving challenges in deep learning-based analyses of histopathological images using explanation methods
title_full_unstemmed Resolving challenges in deep learning-based analyses of histopathological images using explanation methods
title_short Resolving challenges in deep learning-based analyses of histopathological images using explanation methods
title_sort resolving challenges in deep learning-based analyses of histopathological images using explanation methods
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7156509/
https://www.ncbi.nlm.nih.gov/pubmed/32286358
http://dx.doi.org/10.1038/s41598-020-62724-2
work_keys_str_mv AT hagelemiriam resolvingchallengesindeeplearningbasedanalysesofhistopathologicalimagesusingexplanationmethods
AT seegererphilipp resolvingchallengesindeeplearningbasedanalysesofhistopathologicalimagesusingexplanationmethods
AT lapuschkinsebastian resolvingchallengesindeeplearningbasedanalysesofhistopathologicalimagesusingexplanationmethods
AT bockmayrmichael resolvingchallengesindeeplearningbasedanalysesofhistopathologicalimagesusingexplanationmethods
AT samekwojciech resolvingchallengesindeeplearningbasedanalysesofhistopathologicalimagesusingexplanationmethods
AT klauschenfrederick resolvingchallengesindeeplearningbasedanalysesofhistopathologicalimagesusingexplanationmethods
AT mullerklausrobert resolvingchallengesindeeplearningbasedanalysesofhistopathologicalimagesusingexplanationmethods
AT binderalexander resolvingchallengesindeeplearningbasedanalysesofhistopathologicalimagesusingexplanationmethods