Cargando…

Absolute and Relative Depth-Induced Network for RGB-D Salient Object Detection

Detecting salient objects in complicated scenarios is a challenging problem. Except for semantic features from the RGB image, spatial information from the depth image also provides sufficient cues about the object. Therefore, it is crucial to rationally integrate RGB and depth features for the RGB-D...

Descripción completa

Detalles Bibliográficos
Autores principales: Kong, Yuqiu, Wang, He, Kong, Lingwei, Liu, Yang, Yao, Cuili, Yin, Baocai
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10098920/
https://www.ncbi.nlm.nih.gov/pubmed/37050670
http://dx.doi.org/10.3390/s23073611
_version_ 1785024929862778880
author Kong, Yuqiu
Wang, He
Kong, Lingwei
Liu, Yang
Yao, Cuili
Yin, Baocai
author_facet Kong, Yuqiu
Wang, He
Kong, Lingwei
Liu, Yang
Yao, Cuili
Yin, Baocai
author_sort Kong, Yuqiu
collection PubMed
description Detecting salient objects in complicated scenarios is a challenging problem. Except for semantic features from the RGB image, spatial information from the depth image also provides sufficient cues about the object. Therefore, it is crucial to rationally integrate RGB and depth features for the RGB-D salient object detection task. Most existing RGB-D saliency detectors modulate RGB semantic features with absolution depth values. However, they ignore the appearance contrast and structure knowledge indicated by relative depth values between pixels. In this work, we propose a depth-induced network (DIN) for RGB-D salient object detection, to take full advantage of both absolute and relative depth information, and further, enforce the in-depth fusion of the RGB-D cross-modalities. Specifically, an absolute depth-induced module ([Formula: see text]) is proposed, to hierarchically integrate absolute depth values and RGB features, to allow the interaction between the appearance and structural information in the encoding stage. A relative depth-induced module ([Formula: see text]) is designed, to capture detailed saliency cues, by exploring contrastive and structural information from relative depth values in the decoding stage. By combining the [Formula: see text] and [Formula: see text] , we can accurately locate salient objects with clear boundaries, even from complex scenes. The proposed DIN is a lightweight network, and the model size is much smaller than that of state-of-the-art algorithms. Extensive experiments on six challenging benchmarks, show that our method outperforms most existing RGB-D salient object detection models.
format Online
Article
Text
id pubmed-10098920
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-100989202023-04-14 Absolute and Relative Depth-Induced Network for RGB-D Salient Object Detection Kong, Yuqiu Wang, He Kong, Lingwei Liu, Yang Yao, Cuili Yin, Baocai Sensors (Basel) Article Detecting salient objects in complicated scenarios is a challenging problem. Except for semantic features from the RGB image, spatial information from the depth image also provides sufficient cues about the object. Therefore, it is crucial to rationally integrate RGB and depth features for the RGB-D salient object detection task. Most existing RGB-D saliency detectors modulate RGB semantic features with absolution depth values. However, they ignore the appearance contrast and structure knowledge indicated by relative depth values between pixels. In this work, we propose a depth-induced network (DIN) for RGB-D salient object detection, to take full advantage of both absolute and relative depth information, and further, enforce the in-depth fusion of the RGB-D cross-modalities. Specifically, an absolute depth-induced module ([Formula: see text]) is proposed, to hierarchically integrate absolute depth values and RGB features, to allow the interaction between the appearance and structural information in the encoding stage. A relative depth-induced module ([Formula: see text]) is designed, to capture detailed saliency cues, by exploring contrastive and structural information from relative depth values in the decoding stage. By combining the [Formula: see text] and [Formula: see text] , we can accurately locate salient objects with clear boundaries, even from complex scenes. The proposed DIN is a lightweight network, and the model size is much smaller than that of state-of-the-art algorithms. Extensive experiments on six challenging benchmarks, show that our method outperforms most existing RGB-D salient object detection models. MDPI 2023-03-30 /pmc/articles/PMC10098920/ /pubmed/37050670 http://dx.doi.org/10.3390/s23073611 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Kong, Yuqiu
Wang, He
Kong, Lingwei
Liu, Yang
Yao, Cuili
Yin, Baocai
Absolute and Relative Depth-Induced Network for RGB-D Salient Object Detection
title Absolute and Relative Depth-Induced Network for RGB-D Salient Object Detection
title_full Absolute and Relative Depth-Induced Network for RGB-D Salient Object Detection
title_fullStr Absolute and Relative Depth-Induced Network for RGB-D Salient Object Detection
title_full_unstemmed Absolute and Relative Depth-Induced Network for RGB-D Salient Object Detection
title_short Absolute and Relative Depth-Induced Network for RGB-D Salient Object Detection
title_sort absolute and relative depth-induced network for rgb-d salient object detection
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10098920/
https://www.ncbi.nlm.nih.gov/pubmed/37050670
http://dx.doi.org/10.3390/s23073611
work_keys_str_mv AT kongyuqiu absoluteandrelativedepthinducednetworkforrgbdsalientobjectdetection
AT wanghe absoluteandrelativedepthinducednetworkforrgbdsalientobjectdetection
AT konglingwei absoluteandrelativedepthinducednetworkforrgbdsalientobjectdetection
AT liuyang absoluteandrelativedepthinducednetworkforrgbdsalientobjectdetection
AT yaocuili absoluteandrelativedepthinducednetworkforrgbdsalientobjectdetection
AT yinbaocai absoluteandrelativedepthinducednetworkforrgbdsalientobjectdetection