Cargando…

Contrastive learning–guided multi-meta attention network for breast ultrasound video diagnosis

Breast cancer is the most common cause of cancer death in women. Early screening and treatment can effectively improve the success rate of treatment. Ultrasound imaging technology, as the preferred modality for breast cancer screening, provides an essential reference for early diagnosis. Existing co...

Descripción completa

Detalles Bibliográficos
Autores principales: Huang, Xiaoyang, Lin, Zhi, Huang, Shaohui, Wang, Fu Lee, Chan, Moon-Tong, Wang, Liansheng
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9650917/
https://www.ncbi.nlm.nih.gov/pubmed/36387264
http://dx.doi.org/10.3389/fonc.2022.952457
_version_ 1784828129366245376
author Huang, Xiaoyang
Lin, Zhi
Huang, Shaohui
Wang, Fu Lee
Chan, Moon-Tong
Wang, Liansheng
author_facet Huang, Xiaoyang
Lin, Zhi
Huang, Shaohui
Wang, Fu Lee
Chan, Moon-Tong
Wang, Liansheng
author_sort Huang, Xiaoyang
collection PubMed
description Breast cancer is the most common cause of cancer death in women. Early screening and treatment can effectively improve the success rate of treatment. Ultrasound imaging technology, as the preferred modality for breast cancer screening, provides an essential reference for early diagnosis. Existing computer-aided ultrasound imaging diagnostic techniques mainly rely on the selected key frames for breast cancer lesion diagnosis. In this paper, we first collected and annotated a dataset of ultrasound video sequences of 268 cases of breast lesions. Moreover, we propose a contrastive learning–guided multi-meta attention network (CLMAN) by combining a deformed feature extraction module and a multi-meta attention module to address breast lesion diagnosis in ultrasound sequence. The proposed feature extraction module can autonomously acquire key information of the feature map in the spatial dimension, whereas the designed multi-meta attention module is dedicated to effective information aggregation in the temporal dimension. In addition, we utilize a contrast learning strategy to alleviate the problem of high imaging variability within ultrasound lesion videos. The experimental results on our collected dataset show that our CLMAN significantly outperforms existing advanced methods for video classification.
format Online
Article
Text
id pubmed-9650917
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-96509172022-11-15 Contrastive learning–guided multi-meta attention network for breast ultrasound video diagnosis Huang, Xiaoyang Lin, Zhi Huang, Shaohui Wang, Fu Lee Chan, Moon-Tong Wang, Liansheng Front Oncol Oncology Breast cancer is the most common cause of cancer death in women. Early screening and treatment can effectively improve the success rate of treatment. Ultrasound imaging technology, as the preferred modality for breast cancer screening, provides an essential reference for early diagnosis. Existing computer-aided ultrasound imaging diagnostic techniques mainly rely on the selected key frames for breast cancer lesion diagnosis. In this paper, we first collected and annotated a dataset of ultrasound video sequences of 268 cases of breast lesions. Moreover, we propose a contrastive learning–guided multi-meta attention network (CLMAN) by combining a deformed feature extraction module and a multi-meta attention module to address breast lesion diagnosis in ultrasound sequence. The proposed feature extraction module can autonomously acquire key information of the feature map in the spatial dimension, whereas the designed multi-meta attention module is dedicated to effective information aggregation in the temporal dimension. In addition, we utilize a contrast learning strategy to alleviate the problem of high imaging variability within ultrasound lesion videos. The experimental results on our collected dataset show that our CLMAN significantly outperforms existing advanced methods for video classification. Frontiers Media S.A. 2022-10-24 /pmc/articles/PMC9650917/ /pubmed/36387264 http://dx.doi.org/10.3389/fonc.2022.952457 Text en Copyright © 2022 Huang, Lin, Huang, Wang, Chan and Wang https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Oncology
Huang, Xiaoyang
Lin, Zhi
Huang, Shaohui
Wang, Fu Lee
Chan, Moon-Tong
Wang, Liansheng
Contrastive learning–guided multi-meta attention network for breast ultrasound video diagnosis
title Contrastive learning–guided multi-meta attention network for breast ultrasound video diagnosis
title_full Contrastive learning–guided multi-meta attention network for breast ultrasound video diagnosis
title_fullStr Contrastive learning–guided multi-meta attention network for breast ultrasound video diagnosis
title_full_unstemmed Contrastive learning–guided multi-meta attention network for breast ultrasound video diagnosis
title_short Contrastive learning–guided multi-meta attention network for breast ultrasound video diagnosis
title_sort contrastive learning–guided multi-meta attention network for breast ultrasound video diagnosis
topic Oncology
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9650917/
https://www.ncbi.nlm.nih.gov/pubmed/36387264
http://dx.doi.org/10.3389/fonc.2022.952457
work_keys_str_mv AT huangxiaoyang contrastivelearningguidedmultimetaattentionnetworkforbreastultrasoundvideodiagnosis
AT linzhi contrastivelearningguidedmultimetaattentionnetworkforbreastultrasoundvideodiagnosis
AT huangshaohui contrastivelearningguidedmultimetaattentionnetworkforbreastultrasoundvideodiagnosis
AT wangfulee contrastivelearningguidedmultimetaattentionnetworkforbreastultrasoundvideodiagnosis
AT chanmoontong contrastivelearningguidedmultimetaattentionnetworkforbreastultrasoundvideodiagnosis
AT wangliansheng contrastivelearningguidedmultimetaattentionnetworkforbreastultrasoundvideodiagnosis