Cargando…

A Graph-Transformer for Whole Slide Image Classification

Deep learning is a powerful tool for whole slide image (WSI) analysis. Typically, when performing supervised deep learning, a WSI is divided into small patches, trained and the outcomes are aggregated to estimate disease grade. However, patch-based methods introduce label noise during training by as...

Descripción completa

Detalles Bibliográficos
Autores principales: Zheng, Yi, Gindra, Rushin H., Green, Emily J., Burks, Eric J., Betke, Margrit, Beane, Jennifer E., Kolachalama, Vijaya B.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9670036/
https://www.ncbi.nlm.nih.gov/pubmed/35594209
http://dx.doi.org/10.1109/TMI.2022.3176598
_version_ 1784832251293335552
author Zheng, Yi
Gindra, Rushin H.
Green, Emily J.
Burks, Eric J.
Betke, Margrit
Beane, Jennifer E.
Kolachalama, Vijaya B.
author_facet Zheng, Yi
Gindra, Rushin H.
Green, Emily J.
Burks, Eric J.
Betke, Margrit
Beane, Jennifer E.
Kolachalama, Vijaya B.
author_sort Zheng, Yi
collection PubMed
description Deep learning is a powerful tool for whole slide image (WSI) analysis. Typically, when performing supervised deep learning, a WSI is divided into small patches, trained and the outcomes are aggregated to estimate disease grade. However, patch-based methods introduce label noise during training by assuming that each patch is independent with the same label as the WSI and neglect overall WSI-level information that is significant in disease grading. Here we present a Graph-Transformer (GT) that fuses a graph-based representation of an WSI and a vision transformer for processing pathology images, called GTP, to predict disease grade. We selected 4,818 WSIs from the Clinical Proteomic Tumor Analysis Consortium (CPTAC), the National Lung Screening Trial (NLST), and The Cancer Genome Atlas (TCGA), and used GTP to distinguish adenocarcinoma (LUAD) and squamous cell carcinoma (LSCC) from adjacent non-cancerous tissue (normal). First, using NLST data, we developed a contrastive learning framework to generate a feature extractor. This allowed us to compute feature vectors of individual WSI patches, which were used to represent the nodes of the graph followed by construction of the GTP framework. Our model trained on the CPTAC data achieved consistently high performance on three-label classification (normal versus LUAD versus LSCC: mean accuracy = 91.2 ± 2.5%) based on five-fold cross-validation, and mean accuracy = 82.3 ± 1.0% on external test data (TCGA). We also introduced a graph-based saliency mapping technique, called GraphCAM, that can identify regions that are highly associated with the class label. Our findings demonstrate GTP as an interpretable and effective deep learning framework for WSI-level classification.
format Online
Article
Text
id pubmed-9670036
institution National Center for Biotechnology Information
language English
publishDate 2022
record_format MEDLINE/PubMed
spelling pubmed-96700362022-11-17 A Graph-Transformer for Whole Slide Image Classification Zheng, Yi Gindra, Rushin H. Green, Emily J. Burks, Eric J. Betke, Margrit Beane, Jennifer E. Kolachalama, Vijaya B. IEEE Trans Med Imaging Article Deep learning is a powerful tool for whole slide image (WSI) analysis. Typically, when performing supervised deep learning, a WSI is divided into small patches, trained and the outcomes are aggregated to estimate disease grade. However, patch-based methods introduce label noise during training by assuming that each patch is independent with the same label as the WSI and neglect overall WSI-level information that is significant in disease grading. Here we present a Graph-Transformer (GT) that fuses a graph-based representation of an WSI and a vision transformer for processing pathology images, called GTP, to predict disease grade. We selected 4,818 WSIs from the Clinical Proteomic Tumor Analysis Consortium (CPTAC), the National Lung Screening Trial (NLST), and The Cancer Genome Atlas (TCGA), and used GTP to distinguish adenocarcinoma (LUAD) and squamous cell carcinoma (LSCC) from adjacent non-cancerous tissue (normal). First, using NLST data, we developed a contrastive learning framework to generate a feature extractor. This allowed us to compute feature vectors of individual WSI patches, which were used to represent the nodes of the graph followed by construction of the GTP framework. Our model trained on the CPTAC data achieved consistently high performance on three-label classification (normal versus LUAD versus LSCC: mean accuracy = 91.2 ± 2.5%) based on five-fold cross-validation, and mean accuracy = 82.3 ± 1.0% on external test data (TCGA). We also introduced a graph-based saliency mapping technique, called GraphCAM, that can identify regions that are highly associated with the class label. Our findings demonstrate GTP as an interpretable and effective deep learning framework for WSI-level classification. 2022-11 2022-10-27 /pmc/articles/PMC9670036/ /pubmed/35594209 http://dx.doi.org/10.1109/TMI.2022.3176598 Text en https://creativecommons.org/licenses/by/4.0/This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
spellingShingle Article
Zheng, Yi
Gindra, Rushin H.
Green, Emily J.
Burks, Eric J.
Betke, Margrit
Beane, Jennifer E.
Kolachalama, Vijaya B.
A Graph-Transformer for Whole Slide Image Classification
title A Graph-Transformer for Whole Slide Image Classification
title_full A Graph-Transformer for Whole Slide Image Classification
title_fullStr A Graph-Transformer for Whole Slide Image Classification
title_full_unstemmed A Graph-Transformer for Whole Slide Image Classification
title_short A Graph-Transformer for Whole Slide Image Classification
title_sort graph-transformer for whole slide image classification
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9670036/
https://www.ncbi.nlm.nih.gov/pubmed/35594209
http://dx.doi.org/10.1109/TMI.2022.3176598
work_keys_str_mv AT zhengyi agraphtransformerforwholeslideimageclassification
AT gindrarushinh agraphtransformerforwholeslideimageclassification
AT greenemilyj agraphtransformerforwholeslideimageclassification
AT burksericj agraphtransformerforwholeslideimageclassification
AT betkemargrit agraphtransformerforwholeslideimageclassification
AT beanejennifere agraphtransformerforwholeslideimageclassification
AT kolachalamavijayab agraphtransformerforwholeslideimageclassification
AT zhengyi graphtransformerforwholeslideimageclassification
AT gindrarushinh graphtransformerforwholeslideimageclassification
AT greenemilyj graphtransformerforwholeslideimageclassification
AT burksericj graphtransformerforwholeslideimageclassification
AT betkemargrit graphtransformerforwholeslideimageclassification
AT beanejennifere graphtransformerforwholeslideimageclassification
AT kolachalamavijayab graphtransformerforwholeslideimageclassification