Cargando…
CMANet: Cross-Modality Attention Network for Indoor-Scene Semantic Segmentation
Indoor-scene semantic segmentation is of great significance to indoor navigation, high-precision map creation, route planning, etc. However, incorporating RGB and HHA images for indoor-scene semantic segmentation is a promising yet challenging task, due to the diversity of textures and structures an...
Autores principales: | , , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9659145/ https://www.ncbi.nlm.nih.gov/pubmed/36366217 http://dx.doi.org/10.3390/s22218520 |
_version_ | 1784830129120215040 |
---|---|
author | Zhu, Longze Kang, Zhizhong Zhou, Mei Yang, Xi Wang, Zhen Cao, Zhen Ye, Chenming |
author_facet | Zhu, Longze Kang, Zhizhong Zhou, Mei Yang, Xi Wang, Zhen Cao, Zhen Ye, Chenming |
author_sort | Zhu, Longze |
collection | PubMed |
description | Indoor-scene semantic segmentation is of great significance to indoor navigation, high-precision map creation, route planning, etc. However, incorporating RGB and HHA images for indoor-scene semantic segmentation is a promising yet challenging task, due to the diversity of textures and structures and the disparity of multi-modality in physical significance. In this paper, we propose a Cross-Modality Attention Network (CMANet) that facilitates the extraction of both RGB and HHA features and enhances the cross-modality feature integration. CMANet is constructed under the encoder–decoder architecture. The encoder consists of two parallel branches that successively extract the latent modality features from RGB and HHA images, respectively. Particularly, a novel self-attention mechanism-based Cross-Modality Refine Gate (CMRG) is presented, which bridges the two branches. More importantly, the CMRG achieves cross-modality feature fusion and produces certain refined aggregated features; it serves as the most crucial part of CMANet. The decoder is a multi-stage up-sampled backbone that is composed of different residual blocks at each up-sampling stage. Furthermore, bi-directional multi-step propagation and pyramid supervision are applied to assist the leaning process. To evaluate the effectiveness and efficiency of the proposed method, extensive experiments are conducted on NYUDv2 and SUN RGB-D datasets. Experimental results demonstrate that our method outperforms the existing ones for indoor semantic-segmentation tasks. |
format | Online Article Text |
id | pubmed-9659145 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-96591452022-11-15 CMANet: Cross-Modality Attention Network for Indoor-Scene Semantic Segmentation Zhu, Longze Kang, Zhizhong Zhou, Mei Yang, Xi Wang, Zhen Cao, Zhen Ye, Chenming Sensors (Basel) Article Indoor-scene semantic segmentation is of great significance to indoor navigation, high-precision map creation, route planning, etc. However, incorporating RGB and HHA images for indoor-scene semantic segmentation is a promising yet challenging task, due to the diversity of textures and structures and the disparity of multi-modality in physical significance. In this paper, we propose a Cross-Modality Attention Network (CMANet) that facilitates the extraction of both RGB and HHA features and enhances the cross-modality feature integration. CMANet is constructed under the encoder–decoder architecture. The encoder consists of two parallel branches that successively extract the latent modality features from RGB and HHA images, respectively. Particularly, a novel self-attention mechanism-based Cross-Modality Refine Gate (CMRG) is presented, which bridges the two branches. More importantly, the CMRG achieves cross-modality feature fusion and produces certain refined aggregated features; it serves as the most crucial part of CMANet. The decoder is a multi-stage up-sampled backbone that is composed of different residual blocks at each up-sampling stage. Furthermore, bi-directional multi-step propagation and pyramid supervision are applied to assist the leaning process. To evaluate the effectiveness and efficiency of the proposed method, extensive experiments are conducted on NYUDv2 and SUN RGB-D datasets. Experimental results demonstrate that our method outperforms the existing ones for indoor semantic-segmentation tasks. MDPI 2022-11-05 /pmc/articles/PMC9659145/ /pubmed/36366217 http://dx.doi.org/10.3390/s22218520 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Zhu, Longze Kang, Zhizhong Zhou, Mei Yang, Xi Wang, Zhen Cao, Zhen Ye, Chenming CMANet: Cross-Modality Attention Network for Indoor-Scene Semantic Segmentation |
title | CMANet: Cross-Modality Attention Network for Indoor-Scene Semantic Segmentation |
title_full | CMANet: Cross-Modality Attention Network for Indoor-Scene Semantic Segmentation |
title_fullStr | CMANet: Cross-Modality Attention Network for Indoor-Scene Semantic Segmentation |
title_full_unstemmed | CMANet: Cross-Modality Attention Network for Indoor-Scene Semantic Segmentation |
title_short | CMANet: Cross-Modality Attention Network for Indoor-Scene Semantic Segmentation |
title_sort | cmanet: cross-modality attention network for indoor-scene semantic segmentation |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9659145/ https://www.ncbi.nlm.nih.gov/pubmed/36366217 http://dx.doi.org/10.3390/s22218520 |
work_keys_str_mv | AT zhulongze cmanetcrossmodalityattentionnetworkforindoorscenesemanticsegmentation AT kangzhizhong cmanetcrossmodalityattentionnetworkforindoorscenesemanticsegmentation AT zhoumei cmanetcrossmodalityattentionnetworkforindoorscenesemanticsegmentation AT yangxi cmanetcrossmodalityattentionnetworkforindoorscenesemanticsegmentation AT wangzhen cmanetcrossmodalityattentionnetworkforindoorscenesemanticsegmentation AT caozhen cmanetcrossmodalityattentionnetworkforindoorscenesemanticsegmentation AT yechenming cmanetcrossmodalityattentionnetworkforindoorscenesemanticsegmentation |