Cargando…
An Interactive Image Segmentation Method Based on Multi-Level Semantic Fusion
Understanding and analyzing 2D/3D sensor data is crucial for a wide range of machine learning-based applications, including object detection, scene segmentation, and salient object detection. In this context, interactive object segmentation is a vital task in image editing and medical diagnosis, inv...
Autores principales: | , , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10383896/ https://www.ncbi.nlm.nih.gov/pubmed/37514688 http://dx.doi.org/10.3390/s23146394 |
_version_ | 1785081023631982592 |
---|---|
author | Zou, Ruirui Wang, Qinghui Wen, Falin Chen, Yang Liu, Jiale Du, Shaoyi Yuan, Chengzhi |
author_facet | Zou, Ruirui Wang, Qinghui Wen, Falin Chen, Yang Liu, Jiale Du, Shaoyi Yuan, Chengzhi |
author_sort | Zou, Ruirui |
collection | PubMed |
description | Understanding and analyzing 2D/3D sensor data is crucial for a wide range of machine learning-based applications, including object detection, scene segmentation, and salient object detection. In this context, interactive object segmentation is a vital task in image editing and medical diagnosis, involving the accurate separation of the target object from its background based on user annotation information. However, existing interactive object segmentation methods struggle to effectively leverage such information to guide object-segmentation models. To address these challenges, this paper proposes an interactive image-segmentation technique for static images based on multi-level semantic fusion. Our method utilizes user-guidance information both inside and outside the target object to segment it from the static image, making it applicable to both 2D and 3D sensor data. The proposed method introduces a cross-stage feature aggregation module, enabling the effective propagation of multi-scale features from previous stages to the current stage. This mechanism prevents the loss of semantic information caused by multiple upsampling and downsampling of the network, allowing the current stage to make better use of semantic information from the previous stage. Additionally, we incorporate a feature channel attention mechanism to address the issue of rough network segmentation edges. This mechanism captures richer feature details from the feature channel level, leading to finer segmentation edges. In the experimental evaluation conducted on the PASCAL Visual Object Classes (VOC) 2012 dataset, our proposed interactive image segmentation method based on multi-level semantic fusion demonstrates an intersection over union (IOU) accuracy approximately 2.1% higher than the currently popular interactive image segmentation method in static images. The comparative analysis highlights the improved performance and effectiveness of our method. Furthermore, our method exhibits potential applications in various fields, including medical imaging and robotics. Its compatibility with other machine learning methods for visual semantic analysis allows for integration into existing workflows. These aspects emphasize the significance of our contributions in advancing interactive image-segmentation techniques and their practical utility in real-world applications. |
format | Online Article Text |
id | pubmed-10383896 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-103838962023-07-30 An Interactive Image Segmentation Method Based on Multi-Level Semantic Fusion Zou, Ruirui Wang, Qinghui Wen, Falin Chen, Yang Liu, Jiale Du, Shaoyi Yuan, Chengzhi Sensors (Basel) Article Understanding and analyzing 2D/3D sensor data is crucial for a wide range of machine learning-based applications, including object detection, scene segmentation, and salient object detection. In this context, interactive object segmentation is a vital task in image editing and medical diagnosis, involving the accurate separation of the target object from its background based on user annotation information. However, existing interactive object segmentation methods struggle to effectively leverage such information to guide object-segmentation models. To address these challenges, this paper proposes an interactive image-segmentation technique for static images based on multi-level semantic fusion. Our method utilizes user-guidance information both inside and outside the target object to segment it from the static image, making it applicable to both 2D and 3D sensor data. The proposed method introduces a cross-stage feature aggregation module, enabling the effective propagation of multi-scale features from previous stages to the current stage. This mechanism prevents the loss of semantic information caused by multiple upsampling and downsampling of the network, allowing the current stage to make better use of semantic information from the previous stage. Additionally, we incorporate a feature channel attention mechanism to address the issue of rough network segmentation edges. This mechanism captures richer feature details from the feature channel level, leading to finer segmentation edges. In the experimental evaluation conducted on the PASCAL Visual Object Classes (VOC) 2012 dataset, our proposed interactive image segmentation method based on multi-level semantic fusion demonstrates an intersection over union (IOU) accuracy approximately 2.1% higher than the currently popular interactive image segmentation method in static images. The comparative analysis highlights the improved performance and effectiveness of our method. Furthermore, our method exhibits potential applications in various fields, including medical imaging and robotics. Its compatibility with other machine learning methods for visual semantic analysis allows for integration into existing workflows. These aspects emphasize the significance of our contributions in advancing interactive image-segmentation techniques and their practical utility in real-world applications. MDPI 2023-07-14 /pmc/articles/PMC10383896/ /pubmed/37514688 http://dx.doi.org/10.3390/s23146394 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Zou, Ruirui Wang, Qinghui Wen, Falin Chen, Yang Liu, Jiale Du, Shaoyi Yuan, Chengzhi An Interactive Image Segmentation Method Based on Multi-Level Semantic Fusion |
title | An Interactive Image Segmentation Method Based on Multi-Level Semantic Fusion |
title_full | An Interactive Image Segmentation Method Based on Multi-Level Semantic Fusion |
title_fullStr | An Interactive Image Segmentation Method Based on Multi-Level Semantic Fusion |
title_full_unstemmed | An Interactive Image Segmentation Method Based on Multi-Level Semantic Fusion |
title_short | An Interactive Image Segmentation Method Based on Multi-Level Semantic Fusion |
title_sort | interactive image segmentation method based on multi-level semantic fusion |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10383896/ https://www.ncbi.nlm.nih.gov/pubmed/37514688 http://dx.doi.org/10.3390/s23146394 |
work_keys_str_mv | AT zouruirui aninteractiveimagesegmentationmethodbasedonmultilevelsemanticfusion AT wangqinghui aninteractiveimagesegmentationmethodbasedonmultilevelsemanticfusion AT wenfalin aninteractiveimagesegmentationmethodbasedonmultilevelsemanticfusion AT chenyang aninteractiveimagesegmentationmethodbasedonmultilevelsemanticfusion AT liujiale aninteractiveimagesegmentationmethodbasedonmultilevelsemanticfusion AT dushaoyi aninteractiveimagesegmentationmethodbasedonmultilevelsemanticfusion AT yuanchengzhi aninteractiveimagesegmentationmethodbasedonmultilevelsemanticfusion AT zouruirui interactiveimagesegmentationmethodbasedonmultilevelsemanticfusion AT wangqinghui interactiveimagesegmentationmethodbasedonmultilevelsemanticfusion AT wenfalin interactiveimagesegmentationmethodbasedonmultilevelsemanticfusion AT chenyang interactiveimagesegmentationmethodbasedonmultilevelsemanticfusion AT liujiale interactiveimagesegmentationmethodbasedonmultilevelsemanticfusion AT dushaoyi interactiveimagesegmentationmethodbasedonmultilevelsemanticfusion AT yuanchengzhi interactiveimagesegmentationmethodbasedonmultilevelsemanticfusion |