Cargando…

An Interactive Image Segmentation Method Based on Multi-Level Semantic Fusion

Understanding and analyzing 2D/3D sensor data is crucial for a wide range of machine learning-based applications, including object detection, scene segmentation, and salient object detection. In this context, interactive object segmentation is a vital task in image editing and medical diagnosis, inv...

Descripción completa

Detalles Bibliográficos
Autores principales: Zou, Ruirui, Wang, Qinghui, Wen, Falin, Chen, Yang, Liu, Jiale, Du, Shaoyi, Yuan, Chengzhi
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10383896/
https://www.ncbi.nlm.nih.gov/pubmed/37514688
http://dx.doi.org/10.3390/s23146394
_version_ 1785081023631982592
author Zou, Ruirui
Wang, Qinghui
Wen, Falin
Chen, Yang
Liu, Jiale
Du, Shaoyi
Yuan, Chengzhi
author_facet Zou, Ruirui
Wang, Qinghui
Wen, Falin
Chen, Yang
Liu, Jiale
Du, Shaoyi
Yuan, Chengzhi
author_sort Zou, Ruirui
collection PubMed
description Understanding and analyzing 2D/3D sensor data is crucial for a wide range of machine learning-based applications, including object detection, scene segmentation, and salient object detection. In this context, interactive object segmentation is a vital task in image editing and medical diagnosis, involving the accurate separation of the target object from its background based on user annotation information. However, existing interactive object segmentation methods struggle to effectively leverage such information to guide object-segmentation models. To address these challenges, this paper proposes an interactive image-segmentation technique for static images based on multi-level semantic fusion. Our method utilizes user-guidance information both inside and outside the target object to segment it from the static image, making it applicable to both 2D and 3D sensor data. The proposed method introduces a cross-stage feature aggregation module, enabling the effective propagation of multi-scale features from previous stages to the current stage. This mechanism prevents the loss of semantic information caused by multiple upsampling and downsampling of the network, allowing the current stage to make better use of semantic information from the previous stage. Additionally, we incorporate a feature channel attention mechanism to address the issue of rough network segmentation edges. This mechanism captures richer feature details from the feature channel level, leading to finer segmentation edges. In the experimental evaluation conducted on the PASCAL Visual Object Classes (VOC) 2012 dataset, our proposed interactive image segmentation method based on multi-level semantic fusion demonstrates an intersection over union (IOU) accuracy approximately 2.1% higher than the currently popular interactive image segmentation method in static images. The comparative analysis highlights the improved performance and effectiveness of our method. Furthermore, our method exhibits potential applications in various fields, including medical imaging and robotics. Its compatibility with other machine learning methods for visual semantic analysis allows for integration into existing workflows. These aspects emphasize the significance of our contributions in advancing interactive image-segmentation techniques and their practical utility in real-world applications.
format Online
Article
Text
id pubmed-10383896
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-103838962023-07-30 An Interactive Image Segmentation Method Based on Multi-Level Semantic Fusion Zou, Ruirui Wang, Qinghui Wen, Falin Chen, Yang Liu, Jiale Du, Shaoyi Yuan, Chengzhi Sensors (Basel) Article Understanding and analyzing 2D/3D sensor data is crucial for a wide range of machine learning-based applications, including object detection, scene segmentation, and salient object detection. In this context, interactive object segmentation is a vital task in image editing and medical diagnosis, involving the accurate separation of the target object from its background based on user annotation information. However, existing interactive object segmentation methods struggle to effectively leverage such information to guide object-segmentation models. To address these challenges, this paper proposes an interactive image-segmentation technique for static images based on multi-level semantic fusion. Our method utilizes user-guidance information both inside and outside the target object to segment it from the static image, making it applicable to both 2D and 3D sensor data. The proposed method introduces a cross-stage feature aggregation module, enabling the effective propagation of multi-scale features from previous stages to the current stage. This mechanism prevents the loss of semantic information caused by multiple upsampling and downsampling of the network, allowing the current stage to make better use of semantic information from the previous stage. Additionally, we incorporate a feature channel attention mechanism to address the issue of rough network segmentation edges. This mechanism captures richer feature details from the feature channel level, leading to finer segmentation edges. In the experimental evaluation conducted on the PASCAL Visual Object Classes (VOC) 2012 dataset, our proposed interactive image segmentation method based on multi-level semantic fusion demonstrates an intersection over union (IOU) accuracy approximately 2.1% higher than the currently popular interactive image segmentation method in static images. The comparative analysis highlights the improved performance and effectiveness of our method. Furthermore, our method exhibits potential applications in various fields, including medical imaging and robotics. Its compatibility with other machine learning methods for visual semantic analysis allows for integration into existing workflows. These aspects emphasize the significance of our contributions in advancing interactive image-segmentation techniques and their practical utility in real-world applications. MDPI 2023-07-14 /pmc/articles/PMC10383896/ /pubmed/37514688 http://dx.doi.org/10.3390/s23146394 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Zou, Ruirui
Wang, Qinghui
Wen, Falin
Chen, Yang
Liu, Jiale
Du, Shaoyi
Yuan, Chengzhi
An Interactive Image Segmentation Method Based on Multi-Level Semantic Fusion
title An Interactive Image Segmentation Method Based on Multi-Level Semantic Fusion
title_full An Interactive Image Segmentation Method Based on Multi-Level Semantic Fusion
title_fullStr An Interactive Image Segmentation Method Based on Multi-Level Semantic Fusion
title_full_unstemmed An Interactive Image Segmentation Method Based on Multi-Level Semantic Fusion
title_short An Interactive Image Segmentation Method Based on Multi-Level Semantic Fusion
title_sort interactive image segmentation method based on multi-level semantic fusion
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10383896/
https://www.ncbi.nlm.nih.gov/pubmed/37514688
http://dx.doi.org/10.3390/s23146394
work_keys_str_mv AT zouruirui aninteractiveimagesegmentationmethodbasedonmultilevelsemanticfusion
AT wangqinghui aninteractiveimagesegmentationmethodbasedonmultilevelsemanticfusion
AT wenfalin aninteractiveimagesegmentationmethodbasedonmultilevelsemanticfusion
AT chenyang aninteractiveimagesegmentationmethodbasedonmultilevelsemanticfusion
AT liujiale aninteractiveimagesegmentationmethodbasedonmultilevelsemanticfusion
AT dushaoyi aninteractiveimagesegmentationmethodbasedonmultilevelsemanticfusion
AT yuanchengzhi aninteractiveimagesegmentationmethodbasedonmultilevelsemanticfusion
AT zouruirui interactiveimagesegmentationmethodbasedonmultilevelsemanticfusion
AT wangqinghui interactiveimagesegmentationmethodbasedonmultilevelsemanticfusion
AT wenfalin interactiveimagesegmentationmethodbasedonmultilevelsemanticfusion
AT chenyang interactiveimagesegmentationmethodbasedonmultilevelsemanticfusion
AT liujiale interactiveimagesegmentationmethodbasedonmultilevelsemanticfusion
AT dushaoyi interactiveimagesegmentationmethodbasedonmultilevelsemanticfusion
AT yuanchengzhi interactiveimagesegmentationmethodbasedonmultilevelsemanticfusion