Cargando…

Robust fusion for skin lesion segmentation of dermoscopic images

Robust skin lesion segmentation of dermoscopic images is still very difficult. Recent methods often take the combinations of CNN and Transformer for feature abstraction and multi-scale features for further classification. Both types of combination in general rely on some forms of feature fusion. Thi...

Descripción completa

Detalles Bibliográficos
Autores principales: Guo, Qingqing, Fang, Xianyong, Wang, Linbo, Zhang, Enming, Liu, Zhengyi
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10069440/
https://www.ncbi.nlm.nih.gov/pubmed/37020509
http://dx.doi.org/10.3389/fbioe.2023.1057866
_version_ 1785018848685064192
author Guo, Qingqing
Fang, Xianyong
Wang, Linbo
Zhang, Enming
Liu, Zhengyi
author_facet Guo, Qingqing
Fang, Xianyong
Wang, Linbo
Zhang, Enming
Liu, Zhengyi
author_sort Guo, Qingqing
collection PubMed
description Robust skin lesion segmentation of dermoscopic images is still very difficult. Recent methods often take the combinations of CNN and Transformer for feature abstraction and multi-scale features for further classification. Both types of combination in general rely on some forms of feature fusion. This paper considers these fusions from two novel points of view. For abstraction, Transformer is viewed as the affinity exploration of different patch tokens and can be applied to attend CNN features in multiple scales. Consequently, a new fusion module, the Attention-based Transformer-And-CNN fusion module (ATAC), is proposed. ATAC augments the CNN features with more global contexts. For further classification, adaptively combining the information from multiple scales according to their contributions to object recognition is expected. Accordingly, a new fusion module, the GAting-based Multi-Scale fusion module (GAMS), is also introduced, which adaptively weights the information from multiple scales by the light-weighted gating mechanism. Combining ATAC and GAMS leads to a new encoder-decoder-based framework. In this method, ATAC acts as an encoder block to progressively abstract strong CNN features with rich global contexts attended by long-range relations, while GAMS works as an enhancement of the decoder to generate the discriminative features through adaptive fusion of multi-scale ones. This framework is especially good at lesions of varying sizes and shapes and of low contrasts and its performances are demonstrated with extensive experiments on public skin lesion segmentation datasets.
format Online
Article
Text
id pubmed-10069440
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-100694402023-04-04 Robust fusion for skin lesion segmentation of dermoscopic images Guo, Qingqing Fang, Xianyong Wang, Linbo Zhang, Enming Liu, Zhengyi Front Bioeng Biotechnol Bioengineering and Biotechnology Robust skin lesion segmentation of dermoscopic images is still very difficult. Recent methods often take the combinations of CNN and Transformer for feature abstraction and multi-scale features for further classification. Both types of combination in general rely on some forms of feature fusion. This paper considers these fusions from two novel points of view. For abstraction, Transformer is viewed as the affinity exploration of different patch tokens and can be applied to attend CNN features in multiple scales. Consequently, a new fusion module, the Attention-based Transformer-And-CNN fusion module (ATAC), is proposed. ATAC augments the CNN features with more global contexts. For further classification, adaptively combining the information from multiple scales according to their contributions to object recognition is expected. Accordingly, a new fusion module, the GAting-based Multi-Scale fusion module (GAMS), is also introduced, which adaptively weights the information from multiple scales by the light-weighted gating mechanism. Combining ATAC and GAMS leads to a new encoder-decoder-based framework. In this method, ATAC acts as an encoder block to progressively abstract strong CNN features with rich global contexts attended by long-range relations, while GAMS works as an enhancement of the decoder to generate the discriminative features through adaptive fusion of multi-scale ones. This framework is especially good at lesions of varying sizes and shapes and of low contrasts and its performances are demonstrated with extensive experiments on public skin lesion segmentation datasets. Frontiers Media S.A. 2023-03-20 /pmc/articles/PMC10069440/ /pubmed/37020509 http://dx.doi.org/10.3389/fbioe.2023.1057866 Text en Copyright © 2023 Guo, Fang, Wang, Zhang and Liu. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Bioengineering and Biotechnology
Guo, Qingqing
Fang, Xianyong
Wang, Linbo
Zhang, Enming
Liu, Zhengyi
Robust fusion for skin lesion segmentation of dermoscopic images
title Robust fusion for skin lesion segmentation of dermoscopic images
title_full Robust fusion for skin lesion segmentation of dermoscopic images
title_fullStr Robust fusion for skin lesion segmentation of dermoscopic images
title_full_unstemmed Robust fusion for skin lesion segmentation of dermoscopic images
title_short Robust fusion for skin lesion segmentation of dermoscopic images
title_sort robust fusion for skin lesion segmentation of dermoscopic images
topic Bioengineering and Biotechnology
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10069440/
https://www.ncbi.nlm.nih.gov/pubmed/37020509
http://dx.doi.org/10.3389/fbioe.2023.1057866
work_keys_str_mv AT guoqingqing robustfusionforskinlesionsegmentationofdermoscopicimages
AT fangxianyong robustfusionforskinlesionsegmentationofdermoscopicimages
AT wanglinbo robustfusionforskinlesionsegmentationofdermoscopicimages
AT zhangenming robustfusionforskinlesionsegmentationofdermoscopicimages
AT liuzhengyi robustfusionforskinlesionsegmentationofdermoscopicimages