Cargando…
Monocular Depth Estimation: Lightweight Convolutional and Matrix Capsule Feature-Fusion Network
This paper reports a study that aims to solve the problem of the weak adaptability to angle transformation of current monocular depth estimation algorithms. These algorithms are based on convolutional neural networks (CNNs) but produce results lacking in estimation accuracy and robustness. The paper...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9459913/ https://www.ncbi.nlm.nih.gov/pubmed/36080801 http://dx.doi.org/10.3390/s22176344 |
_version_ | 1784786621488431104 |
---|---|
author | Wang, Yinchu Zhu, Haijiang |
author_facet | Wang, Yinchu Zhu, Haijiang |
author_sort | Wang, Yinchu |
collection | PubMed |
description | This paper reports a study that aims to solve the problem of the weak adaptability to angle transformation of current monocular depth estimation algorithms. These algorithms are based on convolutional neural networks (CNNs) but produce results lacking in estimation accuracy and robustness. The paper proposes a lightweight network based on convolution and capsule feature fusion (CNNapsule). First, the paper introduces a fusion block module that integrates CNN features and matrix capsule features to improve the adaptability of the network to perspective transformations. The fusion and deconvolution features are fused through skip connections to generate a depth image. In addition, the corresponding loss function is designed according to the long-tail distribution, gradient similarity, and structural similarity of the datasets. Finally, the results are compared with the methods applied to the NYU Depth V2 and KITTI datasets and show that our proposed method has better accuracy on the C1 and C2 indices and a better visual effect than traditional methods and deep learning methods without transfer learning. The number of trainable parameters required by this method is 65% lower than that required by methods presented in the literature. The generalization of this method is verified via the comparative testing of the data collected from the internet and mobile phones. |
format | Online Article Text |
id | pubmed-9459913 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-94599132022-09-10 Monocular Depth Estimation: Lightweight Convolutional and Matrix Capsule Feature-Fusion Network Wang, Yinchu Zhu, Haijiang Sensors (Basel) Article This paper reports a study that aims to solve the problem of the weak adaptability to angle transformation of current monocular depth estimation algorithms. These algorithms are based on convolutional neural networks (CNNs) but produce results lacking in estimation accuracy and robustness. The paper proposes a lightweight network based on convolution and capsule feature fusion (CNNapsule). First, the paper introduces a fusion block module that integrates CNN features and matrix capsule features to improve the adaptability of the network to perspective transformations. The fusion and deconvolution features are fused through skip connections to generate a depth image. In addition, the corresponding loss function is designed according to the long-tail distribution, gradient similarity, and structural similarity of the datasets. Finally, the results are compared with the methods applied to the NYU Depth V2 and KITTI datasets and show that our proposed method has better accuracy on the C1 and C2 indices and a better visual effect than traditional methods and deep learning methods without transfer learning. The number of trainable parameters required by this method is 65% lower than that required by methods presented in the literature. The generalization of this method is verified via the comparative testing of the data collected from the internet and mobile phones. MDPI 2022-08-23 /pmc/articles/PMC9459913/ /pubmed/36080801 http://dx.doi.org/10.3390/s22176344 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Wang, Yinchu Zhu, Haijiang Monocular Depth Estimation: Lightweight Convolutional and Matrix Capsule Feature-Fusion Network |
title | Monocular Depth Estimation: Lightweight Convolutional and Matrix Capsule Feature-Fusion Network |
title_full | Monocular Depth Estimation: Lightweight Convolutional and Matrix Capsule Feature-Fusion Network |
title_fullStr | Monocular Depth Estimation: Lightweight Convolutional and Matrix Capsule Feature-Fusion Network |
title_full_unstemmed | Monocular Depth Estimation: Lightweight Convolutional and Matrix Capsule Feature-Fusion Network |
title_short | Monocular Depth Estimation: Lightweight Convolutional and Matrix Capsule Feature-Fusion Network |
title_sort | monocular depth estimation: lightweight convolutional and matrix capsule feature-fusion network |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9459913/ https://www.ncbi.nlm.nih.gov/pubmed/36080801 http://dx.doi.org/10.3390/s22176344 |
work_keys_str_mv | AT wangyinchu monoculardepthestimationlightweightconvolutionalandmatrixcapsulefeaturefusionnetwork AT zhuhaijiang monoculardepthestimationlightweightconvolutionalandmatrixcapsulefeaturefusionnetwork |