Cargando…
E(2)LNet: An Efficient and Effective Lightweight Network for Panoramic Depth Estimation
Monocular panoramic depth estimation has various applications in robotics and autonomous driving due to its ability to perceive the entire field of view. However, panoramic depth estimation faces two significant challenges: global context capturing and distortion awareness. In this paper, we propose...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10675273/ https://www.ncbi.nlm.nih.gov/pubmed/38005604 http://dx.doi.org/10.3390/s23229218 |
_version_ | 1785149800478408704 |
---|---|
author | Xu, Jiayue Zhao, Jianping Li, Hua Han, Cheng Xu, Chao |
author_facet | Xu, Jiayue Zhao, Jianping Li, Hua Han, Cheng Xu, Chao |
author_sort | Xu, Jiayue |
collection | PubMed |
description | Monocular panoramic depth estimation has various applications in robotics and autonomous driving due to its ability to perceive the entire field of view. However, panoramic depth estimation faces two significant challenges: global context capturing and distortion awareness. In this paper, we propose a new framework for panoramic depth estimation that can simultaneously address panoramic distortion and extract global context information, thereby improving the performance of panoramic depth estimation. Specifically, we introduce an attention mechanism into the multi-scale dilated convolution and adaptively adjust the receptive field size between different spatial positions, designing the adaptive attention dilated convolution module, which effectively perceives distortion. At the same time, we design the global scene understanding module to integrate global context information into the feature maps generated using the feature extractor. Finally, we trained and evaluated our model on three benchmark datasets which contains the virtual and real-world RGB-D panorama datasets. The experimental results show that the proposed method achieves competitive performance, comparable to existing techniques in both quantitative and qualitative evaluations. Furthermore, our method has fewer parameters and more flexibility, making it a scalable solution in mobile AR. |
format | Online Article Text |
id | pubmed-10675273 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-106752732023-11-16 E(2)LNet: An Efficient and Effective Lightweight Network for Panoramic Depth Estimation Xu, Jiayue Zhao, Jianping Li, Hua Han, Cheng Xu, Chao Sensors (Basel) Article Monocular panoramic depth estimation has various applications in robotics and autonomous driving due to its ability to perceive the entire field of view. However, panoramic depth estimation faces two significant challenges: global context capturing and distortion awareness. In this paper, we propose a new framework for panoramic depth estimation that can simultaneously address panoramic distortion and extract global context information, thereby improving the performance of panoramic depth estimation. Specifically, we introduce an attention mechanism into the multi-scale dilated convolution and adaptively adjust the receptive field size between different spatial positions, designing the adaptive attention dilated convolution module, which effectively perceives distortion. At the same time, we design the global scene understanding module to integrate global context information into the feature maps generated using the feature extractor. Finally, we trained and evaluated our model on three benchmark datasets which contains the virtual and real-world RGB-D panorama datasets. The experimental results show that the proposed method achieves competitive performance, comparable to existing techniques in both quantitative and qualitative evaluations. Furthermore, our method has fewer parameters and more flexibility, making it a scalable solution in mobile AR. MDPI 2023-11-16 /pmc/articles/PMC10675273/ /pubmed/38005604 http://dx.doi.org/10.3390/s23229218 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Xu, Jiayue Zhao, Jianping Li, Hua Han, Cheng Xu, Chao E(2)LNet: An Efficient and Effective Lightweight Network for Panoramic Depth Estimation |
title | E(2)LNet: An Efficient and Effective Lightweight Network for Panoramic Depth Estimation |
title_full | E(2)LNet: An Efficient and Effective Lightweight Network for Panoramic Depth Estimation |
title_fullStr | E(2)LNet: An Efficient and Effective Lightweight Network for Panoramic Depth Estimation |
title_full_unstemmed | E(2)LNet: An Efficient and Effective Lightweight Network for Panoramic Depth Estimation |
title_short | E(2)LNet: An Efficient and Effective Lightweight Network for Panoramic Depth Estimation |
title_sort | e(2)lnet: an efficient and effective lightweight network for panoramic depth estimation |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10675273/ https://www.ncbi.nlm.nih.gov/pubmed/38005604 http://dx.doi.org/10.3390/s23229218 |
work_keys_str_mv | AT xujiayue e2lnetanefficientandeffectivelightweightnetworkforpanoramicdepthestimation AT zhaojianping e2lnetanefficientandeffectivelightweightnetworkforpanoramicdepthestimation AT lihua e2lnetanefficientandeffectivelightweightnetworkforpanoramicdepthestimation AT hancheng e2lnetanefficientandeffectivelightweightnetworkforpanoramicdepthestimation AT xuchao e2lnetanefficientandeffectivelightweightnetworkforpanoramicdepthestimation |