Cargando…

A deep attention LSTM embedded aggregation network for multiple histopathological images

Recent advancements in computer vision and neural networks have facilitated the medical imaging survival analysis for various medical applications. However, challenges arise when patients have multiple images from multiple lesions, as current deep learning methods provide multiple survival predictio...

Descripción completa

Detalles Bibliográficos
Autores principales: Kim, Sunghun, Lee, Eunjee
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Public Library of Science 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10310006/
https://www.ncbi.nlm.nih.gov/pubmed/37384648
http://dx.doi.org/10.1371/journal.pone.0287301
_version_ 1785066494837653504
author Kim, Sunghun
Lee, Eunjee
author_facet Kim, Sunghun
Lee, Eunjee
author_sort Kim, Sunghun
collection PubMed
description Recent advancements in computer vision and neural networks have facilitated the medical imaging survival analysis for various medical applications. However, challenges arise when patients have multiple images from multiple lesions, as current deep learning methods provide multiple survival predictions for each patient, complicating result interpretation. To address this issue, we developed a deep learning survival model that can provide accurate predictions at the patient level. We propose a deep attention long short-term memory embedded aggregation network (DALAN) for histopathology images, designed to simultaneously perform feature extraction and aggregation of lesion images. This design enables the model to efficiently learn imaging features from lesions and aggregate lesion-level information to the patient level. DALAN comprises a weight-shared CNN, attention layers, and LSTM layers. The attention layer calculates the significance of each lesion image, while the LSTM layer combines the weighted information to produce an all-encompassing representation of the patient’s lesion data. Our proposed method performed better on both simulated and real data than other competing methods in terms of prediction accuracy. We evaluated DALAN against several naive aggregation methods on simulated and real datasets. Our results showed that DALAN outperformed the competing methods in terms of c-index on the MNIST and Cancer dataset simulations. On the real TCGA dataset, DALAN also achieved a higher c-index of 0.803±0.006 compared to the naive methods and the competing models. Our DALAN effectively aggregates multiple histopathology images, demonstrating a comprehensive survival model using attention and LSTM mechanisms.
format Online
Article
Text
id pubmed-10310006
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-103100062023-06-30 A deep attention LSTM embedded aggregation network for multiple histopathological images Kim, Sunghun Lee, Eunjee PLoS One Research Article Recent advancements in computer vision and neural networks have facilitated the medical imaging survival analysis for various medical applications. However, challenges arise when patients have multiple images from multiple lesions, as current deep learning methods provide multiple survival predictions for each patient, complicating result interpretation. To address this issue, we developed a deep learning survival model that can provide accurate predictions at the patient level. We propose a deep attention long short-term memory embedded aggregation network (DALAN) for histopathology images, designed to simultaneously perform feature extraction and aggregation of lesion images. This design enables the model to efficiently learn imaging features from lesions and aggregate lesion-level information to the patient level. DALAN comprises a weight-shared CNN, attention layers, and LSTM layers. The attention layer calculates the significance of each lesion image, while the LSTM layer combines the weighted information to produce an all-encompassing representation of the patient’s lesion data. Our proposed method performed better on both simulated and real data than other competing methods in terms of prediction accuracy. We evaluated DALAN against several naive aggregation methods on simulated and real datasets. Our results showed that DALAN outperformed the competing methods in terms of c-index on the MNIST and Cancer dataset simulations. On the real TCGA dataset, DALAN also achieved a higher c-index of 0.803±0.006 compared to the naive methods and the competing models. Our DALAN effectively aggregates multiple histopathology images, demonstrating a comprehensive survival model using attention and LSTM mechanisms. Public Library of Science 2023-06-29 /pmc/articles/PMC10310006/ /pubmed/37384648 http://dx.doi.org/10.1371/journal.pone.0287301 Text en © 2023 Kim, Lee https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
spellingShingle Research Article
Kim, Sunghun
Lee, Eunjee
A deep attention LSTM embedded aggregation network for multiple histopathological images
title A deep attention LSTM embedded aggregation network for multiple histopathological images
title_full A deep attention LSTM embedded aggregation network for multiple histopathological images
title_fullStr A deep attention LSTM embedded aggregation network for multiple histopathological images
title_full_unstemmed A deep attention LSTM embedded aggregation network for multiple histopathological images
title_short A deep attention LSTM embedded aggregation network for multiple histopathological images
title_sort deep attention lstm embedded aggregation network for multiple histopathological images
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10310006/
https://www.ncbi.nlm.nih.gov/pubmed/37384648
http://dx.doi.org/10.1371/journal.pone.0287301
work_keys_str_mv AT kimsunghun adeepattentionlstmembeddedaggregationnetworkformultiplehistopathologicalimages
AT leeeunjee adeepattentionlstmembeddedaggregationnetworkformultiplehistopathologicalimages
AT kimsunghun deepattentionlstmembeddedaggregationnetworkformultiplehistopathologicalimages
AT leeeunjee deepattentionlstmembeddedaggregationnetworkformultiplehistopathologicalimages