Cargando…

Interpretable deep learning approach for oral cancer classification using guided attention inference network

SIGNIFICANCE: Convolutional neural networks (CNNs) show the potential for automated classification of different cancer lesions. However, their lack of interpretability and explainability makes CNNs less than understandable. Furthermore, CNNs may incorrectly concentrate on other areas surrounding the...

Descripción completa

Detalles Bibliográficos
Autores principales: Figueroa, Kevin Chew, Song, Bofan, Sunny, Sumsum, Li, Shaobai, Gurushanth, Keerthi, Mendonca, Pramila, Mukhia, Nirza, Patrick, Sanjana, Gurudath, Shubha, Raghavan, Subhashini, Imchen, Tsusennaro, Leivon, Shirley T., Kolur, Trupti, Shetty, Vivek, Bushan, Vidya, Ramesh, Rohan, Pillai, Vijay, Wilder-Smith, Petra, Sigamani, Alben, Suresh, Amritha, Kuriakose, Moni Abraham, Birur, Praveen, Liang, Rongguang
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Society of Photo-Optical Instrumentation Engineers 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8754153/
https://www.ncbi.nlm.nih.gov/pubmed/35023333
http://dx.doi.org/10.1117/1.JBO.27.1.015001
_version_ 1784632215184867328
author Figueroa, Kevin Chew
Song, Bofan
Sunny, Sumsum
Li, Shaobai
Gurushanth, Keerthi
Mendonca, Pramila
Mukhia, Nirza
Patrick, Sanjana
Gurudath, Shubha
Raghavan, Subhashini
Imchen, Tsusennaro
Leivon, Shirley T.
Kolur, Trupti
Shetty, Vivek
Bushan, Vidya
Ramesh, Rohan
Pillai, Vijay
Wilder-Smith, Petra
Sigamani, Alben
Suresh, Amritha
Kuriakose, Moni Abraham
Birur, Praveen
Liang, Rongguang
author_facet Figueroa, Kevin Chew
Song, Bofan
Sunny, Sumsum
Li, Shaobai
Gurushanth, Keerthi
Mendonca, Pramila
Mukhia, Nirza
Patrick, Sanjana
Gurudath, Shubha
Raghavan, Subhashini
Imchen, Tsusennaro
Leivon, Shirley T.
Kolur, Trupti
Shetty, Vivek
Bushan, Vidya
Ramesh, Rohan
Pillai, Vijay
Wilder-Smith, Petra
Sigamani, Alben
Suresh, Amritha
Kuriakose, Moni Abraham
Birur, Praveen
Liang, Rongguang
author_sort Figueroa, Kevin Chew
collection PubMed
description SIGNIFICANCE: Convolutional neural networks (CNNs) show the potential for automated classification of different cancer lesions. However, their lack of interpretability and explainability makes CNNs less than understandable. Furthermore, CNNs may incorrectly concentrate on other areas surrounding the salient object, rather than the network’s attention focusing directly on the object to be recognized, as the network has no incentive to focus solely on the correct subjects to be detected. This inhibits the reliability of CNNs, especially for biomedical applications. AIM: Develop a deep learning training approach that could provide understandability to its predictions and directly guide the network to concentrate its attention and accurately delineate cancerous regions of the image. APPROACH: We utilized Selvaraju et al.’s gradient-weighted class activation mapping to inject interpretability and explainability into CNNs. We adopted a two-stage training process with data augmentation techniques and Li et al.’s guided attention inference network (GAIN) to train images captured using our customized mobile oral screening devices. The GAIN architecture consists of three streams of network training: classification stream, attention mining stream, and bounding box stream. By adopting the GAIN training architecture, we jointly optimized the classification and segmentation accuracy of our CNN by treating these attention maps as reliable priors to develop attention maps with more complete and accurate segmentation. RESULTS: The network’s attention map will help us to actively understand what the network is focusing on and looking at during its decision-making process. The results also show that the proposed method could guide the trained neural network to highlight and focus its attention on the correct lesion areas in the images when making a decision, rather than focusing its attention on relevant yet incorrect regions. CONCLUSIONS: We demonstrate the effectiveness of our approach for more interpretable and reliable oral potentially malignant lesion and malignant lesion classification.
format Online
Article
Text
id pubmed-8754153
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Society of Photo-Optical Instrumentation Engineers
record_format MEDLINE/PubMed
spelling pubmed-87541532022-01-13 Interpretable deep learning approach for oral cancer classification using guided attention inference network Figueroa, Kevin Chew Song, Bofan Sunny, Sumsum Li, Shaobai Gurushanth, Keerthi Mendonca, Pramila Mukhia, Nirza Patrick, Sanjana Gurudath, Shubha Raghavan, Subhashini Imchen, Tsusennaro Leivon, Shirley T. Kolur, Trupti Shetty, Vivek Bushan, Vidya Ramesh, Rohan Pillai, Vijay Wilder-Smith, Petra Sigamani, Alben Suresh, Amritha Kuriakose, Moni Abraham Birur, Praveen Liang, Rongguang J Biomed Opt General SIGNIFICANCE: Convolutional neural networks (CNNs) show the potential for automated classification of different cancer lesions. However, their lack of interpretability and explainability makes CNNs less than understandable. Furthermore, CNNs may incorrectly concentrate on other areas surrounding the salient object, rather than the network’s attention focusing directly on the object to be recognized, as the network has no incentive to focus solely on the correct subjects to be detected. This inhibits the reliability of CNNs, especially for biomedical applications. AIM: Develop a deep learning training approach that could provide understandability to its predictions and directly guide the network to concentrate its attention and accurately delineate cancerous regions of the image. APPROACH: We utilized Selvaraju et al.’s gradient-weighted class activation mapping to inject interpretability and explainability into CNNs. We adopted a two-stage training process with data augmentation techniques and Li et al.’s guided attention inference network (GAIN) to train images captured using our customized mobile oral screening devices. The GAIN architecture consists of three streams of network training: classification stream, attention mining stream, and bounding box stream. By adopting the GAIN training architecture, we jointly optimized the classification and segmentation accuracy of our CNN by treating these attention maps as reliable priors to develop attention maps with more complete and accurate segmentation. RESULTS: The network’s attention map will help us to actively understand what the network is focusing on and looking at during its decision-making process. The results also show that the proposed method could guide the trained neural network to highlight and focus its attention on the correct lesion areas in the images when making a decision, rather than focusing its attention on relevant yet incorrect regions. CONCLUSIONS: We demonstrate the effectiveness of our approach for more interpretable and reliable oral potentially malignant lesion and malignant lesion classification. Society of Photo-Optical Instrumentation Engineers 2022-01-12 2022-01 /pmc/articles/PMC8754153/ /pubmed/35023333 http://dx.doi.org/10.1117/1.JBO.27.1.015001 Text en © 2022 The Authors https://creativecommons.org/licenses/by/4.0/Published by SPIE under a Creative Commons Attribution 4.0 International License. Distribution or reproduction of this work in whole or in part requires full attribution of the original publication, including its DOI.
spellingShingle General
Figueroa, Kevin Chew
Song, Bofan
Sunny, Sumsum
Li, Shaobai
Gurushanth, Keerthi
Mendonca, Pramila
Mukhia, Nirza
Patrick, Sanjana
Gurudath, Shubha
Raghavan, Subhashini
Imchen, Tsusennaro
Leivon, Shirley T.
Kolur, Trupti
Shetty, Vivek
Bushan, Vidya
Ramesh, Rohan
Pillai, Vijay
Wilder-Smith, Petra
Sigamani, Alben
Suresh, Amritha
Kuriakose, Moni Abraham
Birur, Praveen
Liang, Rongguang
Interpretable deep learning approach for oral cancer classification using guided attention inference network
title Interpretable deep learning approach for oral cancer classification using guided attention inference network
title_full Interpretable deep learning approach for oral cancer classification using guided attention inference network
title_fullStr Interpretable deep learning approach for oral cancer classification using guided attention inference network
title_full_unstemmed Interpretable deep learning approach for oral cancer classification using guided attention inference network
title_short Interpretable deep learning approach for oral cancer classification using guided attention inference network
title_sort interpretable deep learning approach for oral cancer classification using guided attention inference network
topic General
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8754153/
https://www.ncbi.nlm.nih.gov/pubmed/35023333
http://dx.doi.org/10.1117/1.JBO.27.1.015001
work_keys_str_mv AT figueroakevinchew interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork
AT songbofan interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork
AT sunnysumsum interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork
AT lishaobai interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork
AT gurushanthkeerthi interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork
AT mendoncapramila interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork
AT mukhianirza interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork
AT patricksanjana interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork
AT gurudathshubha interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork
AT raghavansubhashini interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork
AT imchentsusennaro interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork
AT leivonshirleyt interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork
AT kolurtrupti interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork
AT shettyvivek interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork
AT bushanvidya interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork
AT rameshrohan interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork
AT pillaivijay interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork
AT wildersmithpetra interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork
AT sigamanialben interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork
AT sureshamritha interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork
AT kuriakosemoniabraham interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork
AT birurpraveen interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork
AT liangrongguang interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork