Cargando…

Anti-Aliasing Attention U-net Model for Skin Lesion Segmentation

The need for a lightweight and reliable segmentation algorithm is critical in various biomedical image-prediction applications. However, the limited quantity of data presents a significant challenge for image segmentation. Additionally, low image quality negatively impacts the efficiency of segmenta...

Descripción completa

Detalles Bibliográficos
Autores principales: Le, Phuong Thi, Pham, Bach-Tung, Chang, Ching-Chun, Hsu, Yi-Chiung, Tai, Tzu-Chiang, Li, Yung-Hui, Wang, Jia-Ching
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10137492/
https://www.ncbi.nlm.nih.gov/pubmed/37189563
http://dx.doi.org/10.3390/diagnostics13081460
_version_ 1785032476684451840
author Le, Phuong Thi
Pham, Bach-Tung
Chang, Ching-Chun
Hsu, Yi-Chiung
Tai, Tzu-Chiang
Li, Yung-Hui
Wang, Jia-Ching
author_facet Le, Phuong Thi
Pham, Bach-Tung
Chang, Ching-Chun
Hsu, Yi-Chiung
Tai, Tzu-Chiang
Li, Yung-Hui
Wang, Jia-Ching
author_sort Le, Phuong Thi
collection PubMed
description The need for a lightweight and reliable segmentation algorithm is critical in various biomedical image-prediction applications. However, the limited quantity of data presents a significant challenge for image segmentation. Additionally, low image quality negatively impacts the efficiency of segmentation, and previous deep learning models for image segmentation require large parameters with hundreds of millions of computations, resulting in high costs and processing times. In this study, we introduce a new lightweight segmentation model, the mobile anti-aliasing attention u-net model (MAAU), which features both encoder and decoder paths. The encoder incorporates an anti-aliasing layer and convolutional blocks to reduce the spatial resolution of input images while avoiding shift equivariance. The decoder uses an attention block and decoder module to capture prominent features in each channel. To address data-related problems, we implemented data augmentation methods such as flip, rotation, shear, translate, and color distortions, which enhanced segmentation efficiency in the international Skin Image Collaboration (ISIC) 2018 and PH2 datasets. Our experimental results demonstrated that our approach had fewer parameters, only 4.2 million, while it outperformed various state-of-the-art segmentation methods.
format Online
Article
Text
id pubmed-10137492
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-101374922023-04-28 Anti-Aliasing Attention U-net Model for Skin Lesion Segmentation Le, Phuong Thi Pham, Bach-Tung Chang, Ching-Chun Hsu, Yi-Chiung Tai, Tzu-Chiang Li, Yung-Hui Wang, Jia-Ching Diagnostics (Basel) Article The need for a lightweight and reliable segmentation algorithm is critical in various biomedical image-prediction applications. However, the limited quantity of data presents a significant challenge for image segmentation. Additionally, low image quality negatively impacts the efficiency of segmentation, and previous deep learning models for image segmentation require large parameters with hundreds of millions of computations, resulting in high costs and processing times. In this study, we introduce a new lightweight segmentation model, the mobile anti-aliasing attention u-net model (MAAU), which features both encoder and decoder paths. The encoder incorporates an anti-aliasing layer and convolutional blocks to reduce the spatial resolution of input images while avoiding shift equivariance. The decoder uses an attention block and decoder module to capture prominent features in each channel. To address data-related problems, we implemented data augmentation methods such as flip, rotation, shear, translate, and color distortions, which enhanced segmentation efficiency in the international Skin Image Collaboration (ISIC) 2018 and PH2 datasets. Our experimental results demonstrated that our approach had fewer parameters, only 4.2 million, while it outperformed various state-of-the-art segmentation methods. MDPI 2023-04-18 /pmc/articles/PMC10137492/ /pubmed/37189563 http://dx.doi.org/10.3390/diagnostics13081460 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Le, Phuong Thi
Pham, Bach-Tung
Chang, Ching-Chun
Hsu, Yi-Chiung
Tai, Tzu-Chiang
Li, Yung-Hui
Wang, Jia-Ching
Anti-Aliasing Attention U-net Model for Skin Lesion Segmentation
title Anti-Aliasing Attention U-net Model for Skin Lesion Segmentation
title_full Anti-Aliasing Attention U-net Model for Skin Lesion Segmentation
title_fullStr Anti-Aliasing Attention U-net Model for Skin Lesion Segmentation
title_full_unstemmed Anti-Aliasing Attention U-net Model for Skin Lesion Segmentation
title_short Anti-Aliasing Attention U-net Model for Skin Lesion Segmentation
title_sort anti-aliasing attention u-net model for skin lesion segmentation
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10137492/
https://www.ncbi.nlm.nih.gov/pubmed/37189563
http://dx.doi.org/10.3390/diagnostics13081460
work_keys_str_mv AT lephuongthi antialiasingattentionunetmodelforskinlesionsegmentation
AT phambachtung antialiasingattentionunetmodelforskinlesionsegmentation
AT changchingchun antialiasingattentionunetmodelforskinlesionsegmentation
AT hsuyichiung antialiasingattentionunetmodelforskinlesionsegmentation
AT taitzuchiang antialiasingattentionunetmodelforskinlesionsegmentation
AT liyunghui antialiasingattentionunetmodelforskinlesionsegmentation
AT wangjiaching antialiasingattentionunetmodelforskinlesionsegmentation