Cargando…

SU(2)GE-Net: a saliency-based approach for non-specific class foreground segmentation

Salient object detection is vital for non-specific class subject segmentation in computer vision applications. However, accurately segmenting foreground subjects with complex backgrounds and intricate boundaries remains a challenge for existing methods. To address these limitations, our study propos...

Descripción completa

Detalles Bibliográficos
Autores principales: Lei, Xiaochun, Cai, Xiang, Lu, Linjun, Cui, Zihang, Jiang, Zetao
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Nature Publishing Group UK 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10427708/
https://www.ncbi.nlm.nih.gov/pubmed/37582948
http://dx.doi.org/10.1038/s41598-023-40175-9
Descripción
Sumario:Salient object detection is vital for non-specific class subject segmentation in computer vision applications. However, accurately segmenting foreground subjects with complex backgrounds and intricate boundaries remains a challenge for existing methods. To address these limitations, our study proposes SU(2)GE-Net, which introduces several novel improvements. We replace the traditional CNN-based backbone with the transformer-based Swin-TransformerV2, known for its effectiveness in capturing long-range dependencies and rich contextual information. To tackle under and over-attention phenomena, we introduce Gated Channel Transformation (GCT). Furthermore, we adopted an edge-based loss (Edge Loss) for network training to capture spatial-wise structural details. Additionally, we propose Training-only Augmentation Loss (TTA Loss) to enhance spatial stability using augmented data. Our method is evaluated using six common datasets, achieving an impressive [Formula: see text] score of 0.883 on DUTS-TE. Compared with other models, SU(2)GE-Net demonstrates excellent performance in various segmentation scenarios.