Cargando…
Interpretable deep learning approach for oral cancer classification using guided attention inference network
SIGNIFICANCE: Convolutional neural networks (CNNs) show the potential for automated classification of different cancer lesions. However, their lack of interpretability and explainability makes CNNs less than understandable. Furthermore, CNNs may incorrectly concentrate on other areas surrounding the...
Autores principales: | , , , , , , , , , , , , , , , , , , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Society of Photo-Optical Instrumentation Engineers
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8754153/ https://www.ncbi.nlm.nih.gov/pubmed/35023333 http://dx.doi.org/10.1117/1.JBO.27.1.015001 |
_version_ | 1784632215184867328 |
---|---|
author | Figueroa, Kevin Chew Song, Bofan Sunny, Sumsum Li, Shaobai Gurushanth, Keerthi Mendonca, Pramila Mukhia, Nirza Patrick, Sanjana Gurudath, Shubha Raghavan, Subhashini Imchen, Tsusennaro Leivon, Shirley T. Kolur, Trupti Shetty, Vivek Bushan, Vidya Ramesh, Rohan Pillai, Vijay Wilder-Smith, Petra Sigamani, Alben Suresh, Amritha Kuriakose, Moni Abraham Birur, Praveen Liang, Rongguang |
author_facet | Figueroa, Kevin Chew Song, Bofan Sunny, Sumsum Li, Shaobai Gurushanth, Keerthi Mendonca, Pramila Mukhia, Nirza Patrick, Sanjana Gurudath, Shubha Raghavan, Subhashini Imchen, Tsusennaro Leivon, Shirley T. Kolur, Trupti Shetty, Vivek Bushan, Vidya Ramesh, Rohan Pillai, Vijay Wilder-Smith, Petra Sigamani, Alben Suresh, Amritha Kuriakose, Moni Abraham Birur, Praveen Liang, Rongguang |
author_sort | Figueroa, Kevin Chew |
collection | PubMed |
description | SIGNIFICANCE: Convolutional neural networks (CNNs) show the potential for automated classification of different cancer lesions. However, their lack of interpretability and explainability makes CNNs less than understandable. Furthermore, CNNs may incorrectly concentrate on other areas surrounding the salient object, rather than the network’s attention focusing directly on the object to be recognized, as the network has no incentive to focus solely on the correct subjects to be detected. This inhibits the reliability of CNNs, especially for biomedical applications. AIM: Develop a deep learning training approach that could provide understandability to its predictions and directly guide the network to concentrate its attention and accurately delineate cancerous regions of the image. APPROACH: We utilized Selvaraju et al.’s gradient-weighted class activation mapping to inject interpretability and explainability into CNNs. We adopted a two-stage training process with data augmentation techniques and Li et al.’s guided attention inference network (GAIN) to train images captured using our customized mobile oral screening devices. The GAIN architecture consists of three streams of network training: classification stream, attention mining stream, and bounding box stream. By adopting the GAIN training architecture, we jointly optimized the classification and segmentation accuracy of our CNN by treating these attention maps as reliable priors to develop attention maps with more complete and accurate segmentation. RESULTS: The network’s attention map will help us to actively understand what the network is focusing on and looking at during its decision-making process. The results also show that the proposed method could guide the trained neural network to highlight and focus its attention on the correct lesion areas in the images when making a decision, rather than focusing its attention on relevant yet incorrect regions. CONCLUSIONS: We demonstrate the effectiveness of our approach for more interpretable and reliable oral potentially malignant lesion and malignant lesion classification. |
format | Online Article Text |
id | pubmed-8754153 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Society of Photo-Optical Instrumentation Engineers |
record_format | MEDLINE/PubMed |
spelling | pubmed-87541532022-01-13 Interpretable deep learning approach for oral cancer classification using guided attention inference network Figueroa, Kevin Chew Song, Bofan Sunny, Sumsum Li, Shaobai Gurushanth, Keerthi Mendonca, Pramila Mukhia, Nirza Patrick, Sanjana Gurudath, Shubha Raghavan, Subhashini Imchen, Tsusennaro Leivon, Shirley T. Kolur, Trupti Shetty, Vivek Bushan, Vidya Ramesh, Rohan Pillai, Vijay Wilder-Smith, Petra Sigamani, Alben Suresh, Amritha Kuriakose, Moni Abraham Birur, Praveen Liang, Rongguang J Biomed Opt General SIGNIFICANCE: Convolutional neural networks (CNNs) show the potential for automated classification of different cancer lesions. However, their lack of interpretability and explainability makes CNNs less than understandable. Furthermore, CNNs may incorrectly concentrate on other areas surrounding the salient object, rather than the network’s attention focusing directly on the object to be recognized, as the network has no incentive to focus solely on the correct subjects to be detected. This inhibits the reliability of CNNs, especially for biomedical applications. AIM: Develop a deep learning training approach that could provide understandability to its predictions and directly guide the network to concentrate its attention and accurately delineate cancerous regions of the image. APPROACH: We utilized Selvaraju et al.’s gradient-weighted class activation mapping to inject interpretability and explainability into CNNs. We adopted a two-stage training process with data augmentation techniques and Li et al.’s guided attention inference network (GAIN) to train images captured using our customized mobile oral screening devices. The GAIN architecture consists of three streams of network training: classification stream, attention mining stream, and bounding box stream. By adopting the GAIN training architecture, we jointly optimized the classification and segmentation accuracy of our CNN by treating these attention maps as reliable priors to develop attention maps with more complete and accurate segmentation. RESULTS: The network’s attention map will help us to actively understand what the network is focusing on and looking at during its decision-making process. The results also show that the proposed method could guide the trained neural network to highlight and focus its attention on the correct lesion areas in the images when making a decision, rather than focusing its attention on relevant yet incorrect regions. CONCLUSIONS: We demonstrate the effectiveness of our approach for more interpretable and reliable oral potentially malignant lesion and malignant lesion classification. Society of Photo-Optical Instrumentation Engineers 2022-01-12 2022-01 /pmc/articles/PMC8754153/ /pubmed/35023333 http://dx.doi.org/10.1117/1.JBO.27.1.015001 Text en © 2022 The Authors https://creativecommons.org/licenses/by/4.0/Published by SPIE under a Creative Commons Attribution 4.0 International License. Distribution or reproduction of this work in whole or in part requires full attribution of the original publication, including its DOI. |
spellingShingle | General Figueroa, Kevin Chew Song, Bofan Sunny, Sumsum Li, Shaobai Gurushanth, Keerthi Mendonca, Pramila Mukhia, Nirza Patrick, Sanjana Gurudath, Shubha Raghavan, Subhashini Imchen, Tsusennaro Leivon, Shirley T. Kolur, Trupti Shetty, Vivek Bushan, Vidya Ramesh, Rohan Pillai, Vijay Wilder-Smith, Petra Sigamani, Alben Suresh, Amritha Kuriakose, Moni Abraham Birur, Praveen Liang, Rongguang Interpretable deep learning approach for oral cancer classification using guided attention inference network |
title | Interpretable deep learning approach for oral cancer classification using guided attention inference network |
title_full | Interpretable deep learning approach for oral cancer classification using guided attention inference network |
title_fullStr | Interpretable deep learning approach for oral cancer classification using guided attention inference network |
title_full_unstemmed | Interpretable deep learning approach for oral cancer classification using guided attention inference network |
title_short | Interpretable deep learning approach for oral cancer classification using guided attention inference network |
title_sort | interpretable deep learning approach for oral cancer classification using guided attention inference network |
topic | General |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8754153/ https://www.ncbi.nlm.nih.gov/pubmed/35023333 http://dx.doi.org/10.1117/1.JBO.27.1.015001 |
work_keys_str_mv | AT figueroakevinchew interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork AT songbofan interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork AT sunnysumsum interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork AT lishaobai interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork AT gurushanthkeerthi interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork AT mendoncapramila interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork AT mukhianirza interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork AT patricksanjana interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork AT gurudathshubha interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork AT raghavansubhashini interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork AT imchentsusennaro interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork AT leivonshirleyt interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork AT kolurtrupti interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork AT shettyvivek interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork AT bushanvidya interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork AT rameshrohan interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork AT pillaivijay interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork AT wildersmithpetra interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork AT sigamanialben interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork AT sureshamritha interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork AT kuriakosemoniabraham interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork AT birurpraveen interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork AT liangrongguang interpretabledeeplearningapproachfororalcancerclassificationusingguidedattentioninferencenetwork |