Cargando…
Absolute and Relative Depth-Induced Network for RGB-D Salient Object Detection
Detecting salient objects in complicated scenarios is a challenging problem. Except for semantic features from the RGB image, spatial information from the depth image also provides sufficient cues about the object. Therefore, it is crucial to rationally integrate RGB and depth features for the RGB-D...
Autores principales: | , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10098920/ https://www.ncbi.nlm.nih.gov/pubmed/37050670 http://dx.doi.org/10.3390/s23073611 |
_version_ | 1785024929862778880 |
---|---|
author | Kong, Yuqiu Wang, He Kong, Lingwei Liu, Yang Yao, Cuili Yin, Baocai |
author_facet | Kong, Yuqiu Wang, He Kong, Lingwei Liu, Yang Yao, Cuili Yin, Baocai |
author_sort | Kong, Yuqiu |
collection | PubMed |
description | Detecting salient objects in complicated scenarios is a challenging problem. Except for semantic features from the RGB image, spatial information from the depth image also provides sufficient cues about the object. Therefore, it is crucial to rationally integrate RGB and depth features for the RGB-D salient object detection task. Most existing RGB-D saliency detectors modulate RGB semantic features with absolution depth values. However, they ignore the appearance contrast and structure knowledge indicated by relative depth values between pixels. In this work, we propose a depth-induced network (DIN) for RGB-D salient object detection, to take full advantage of both absolute and relative depth information, and further, enforce the in-depth fusion of the RGB-D cross-modalities. Specifically, an absolute depth-induced module ([Formula: see text]) is proposed, to hierarchically integrate absolute depth values and RGB features, to allow the interaction between the appearance and structural information in the encoding stage. A relative depth-induced module ([Formula: see text]) is designed, to capture detailed saliency cues, by exploring contrastive and structural information from relative depth values in the decoding stage. By combining the [Formula: see text] and [Formula: see text] , we can accurately locate salient objects with clear boundaries, even from complex scenes. The proposed DIN is a lightweight network, and the model size is much smaller than that of state-of-the-art algorithms. Extensive experiments on six challenging benchmarks, show that our method outperforms most existing RGB-D salient object detection models. |
format | Online Article Text |
id | pubmed-10098920 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-100989202023-04-14 Absolute and Relative Depth-Induced Network for RGB-D Salient Object Detection Kong, Yuqiu Wang, He Kong, Lingwei Liu, Yang Yao, Cuili Yin, Baocai Sensors (Basel) Article Detecting salient objects in complicated scenarios is a challenging problem. Except for semantic features from the RGB image, spatial information from the depth image also provides sufficient cues about the object. Therefore, it is crucial to rationally integrate RGB and depth features for the RGB-D salient object detection task. Most existing RGB-D saliency detectors modulate RGB semantic features with absolution depth values. However, they ignore the appearance contrast and structure knowledge indicated by relative depth values between pixels. In this work, we propose a depth-induced network (DIN) for RGB-D salient object detection, to take full advantage of both absolute and relative depth information, and further, enforce the in-depth fusion of the RGB-D cross-modalities. Specifically, an absolute depth-induced module ([Formula: see text]) is proposed, to hierarchically integrate absolute depth values and RGB features, to allow the interaction between the appearance and structural information in the encoding stage. A relative depth-induced module ([Formula: see text]) is designed, to capture detailed saliency cues, by exploring contrastive and structural information from relative depth values in the decoding stage. By combining the [Formula: see text] and [Formula: see text] , we can accurately locate salient objects with clear boundaries, even from complex scenes. The proposed DIN is a lightweight network, and the model size is much smaller than that of state-of-the-art algorithms. Extensive experiments on six challenging benchmarks, show that our method outperforms most existing RGB-D salient object detection models. MDPI 2023-03-30 /pmc/articles/PMC10098920/ /pubmed/37050670 http://dx.doi.org/10.3390/s23073611 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Kong, Yuqiu Wang, He Kong, Lingwei Liu, Yang Yao, Cuili Yin, Baocai Absolute and Relative Depth-Induced Network for RGB-D Salient Object Detection |
title | Absolute and Relative Depth-Induced Network for RGB-D Salient Object Detection |
title_full | Absolute and Relative Depth-Induced Network for RGB-D Salient Object Detection |
title_fullStr | Absolute and Relative Depth-Induced Network for RGB-D Salient Object Detection |
title_full_unstemmed | Absolute and Relative Depth-Induced Network for RGB-D Salient Object Detection |
title_short | Absolute and Relative Depth-Induced Network for RGB-D Salient Object Detection |
title_sort | absolute and relative depth-induced network for rgb-d salient object detection |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10098920/ https://www.ncbi.nlm.nih.gov/pubmed/37050670 http://dx.doi.org/10.3390/s23073611 |
work_keys_str_mv | AT kongyuqiu absoluteandrelativedepthinducednetworkforrgbdsalientobjectdetection AT wanghe absoluteandrelativedepthinducednetworkforrgbdsalientobjectdetection AT konglingwei absoluteandrelativedepthinducednetworkforrgbdsalientobjectdetection AT liuyang absoluteandrelativedepthinducednetworkforrgbdsalientobjectdetection AT yaocuili absoluteandrelativedepthinducednetworkforrgbdsalientobjectdetection AT yinbaocai absoluteandrelativedepthinducednetworkforrgbdsalientobjectdetection |