Cargando…

DRI-MVSNet: A depth residual inference network for multi-view stereo images

Three-dimensional (3D) image reconstruction is an important field of computer vision for restoring the 3D geometry of a given scene. Due to the demand for large amounts of memory, prevalent methods of 3D reconstruction yield inaccurate results, because of which the highly accuracy reconstruction of...

Descripción completa

Detalles Bibliográficos
Autores principales: Li, Ying, Li, Wenyue, Zhao, Zhijie, Fan, JiaHao
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Public Library of Science 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8942269/
https://www.ncbi.nlm.nih.gov/pubmed/35320265
http://dx.doi.org/10.1371/journal.pone.0264721
_version_ 1784673270298050560
author Li, Ying
Li, Wenyue
Zhao, Zhijie
Fan, JiaHao
author_facet Li, Ying
Li, Wenyue
Zhao, Zhijie
Fan, JiaHao
author_sort Li, Ying
collection PubMed
description Three-dimensional (3D) image reconstruction is an important field of computer vision for restoring the 3D geometry of a given scene. Due to the demand for large amounts of memory, prevalent methods of 3D reconstruction yield inaccurate results, because of which the highly accuracy reconstruction of a scene remains an outstanding challenge. This study proposes a cascaded depth residual inference network, called DRI-MVSNet, that uses a cross-view similarity-based feature map fusion module for residual inference. It involves three improvements. First, a combined module is used for processing channel-related and spatial information to capture the relevant contextual information and improve feature representation. It combines the channel attention mechanism and spatial pooling networks. Second, a cross-view similarity-based feature map fusion module is proposed that learns the similarity between pairs of pixel in each source and reference image at planes of different depths along the frustum of the reference camera. Third, a deep, multi-stage residual prediction module is designed to generate a high-precision depth map that uses a non-uniform depth sampling strategy to construct hypothetical depth planes. The results of extensive experiments show that DRI-MVSNet delivers competitive performance on the DTU and the Tanks & Temples datasets, and the accuracy and completeness of the point cloud reconstructed by it are significantly superior to those of state-of-the-art benchmarks.
format Online
Article
Text
id pubmed-8942269
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-89422692022-03-24 DRI-MVSNet: A depth residual inference network for multi-view stereo images Li, Ying Li, Wenyue Zhao, Zhijie Fan, JiaHao PLoS One Research Article Three-dimensional (3D) image reconstruction is an important field of computer vision for restoring the 3D geometry of a given scene. Due to the demand for large amounts of memory, prevalent methods of 3D reconstruction yield inaccurate results, because of which the highly accuracy reconstruction of a scene remains an outstanding challenge. This study proposes a cascaded depth residual inference network, called DRI-MVSNet, that uses a cross-view similarity-based feature map fusion module for residual inference. It involves three improvements. First, a combined module is used for processing channel-related and spatial information to capture the relevant contextual information and improve feature representation. It combines the channel attention mechanism and spatial pooling networks. Second, a cross-view similarity-based feature map fusion module is proposed that learns the similarity between pairs of pixel in each source and reference image at planes of different depths along the frustum of the reference camera. Third, a deep, multi-stage residual prediction module is designed to generate a high-precision depth map that uses a non-uniform depth sampling strategy to construct hypothetical depth planes. The results of extensive experiments show that DRI-MVSNet delivers competitive performance on the DTU and the Tanks & Temples datasets, and the accuracy and completeness of the point cloud reconstructed by it are significantly superior to those of state-of-the-art benchmarks. Public Library of Science 2022-03-23 /pmc/articles/PMC8942269/ /pubmed/35320265 http://dx.doi.org/10.1371/journal.pone.0264721 Text en © 2022 Li et al https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
spellingShingle Research Article
Li, Ying
Li, Wenyue
Zhao, Zhijie
Fan, JiaHao
DRI-MVSNet: A depth residual inference network for multi-view stereo images
title DRI-MVSNet: A depth residual inference network for multi-view stereo images
title_full DRI-MVSNet: A depth residual inference network for multi-view stereo images
title_fullStr DRI-MVSNet: A depth residual inference network for multi-view stereo images
title_full_unstemmed DRI-MVSNet: A depth residual inference network for multi-view stereo images
title_short DRI-MVSNet: A depth residual inference network for multi-view stereo images
title_sort dri-mvsnet: a depth residual inference network for multi-view stereo images
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8942269/
https://www.ncbi.nlm.nih.gov/pubmed/35320265
http://dx.doi.org/10.1371/journal.pone.0264721
work_keys_str_mv AT liying drimvsnetadepthresidualinferencenetworkformultiviewstereoimages
AT liwenyue drimvsnetadepthresidualinferencenetworkformultiviewstereoimages
AT zhaozhijie drimvsnetadepthresidualinferencenetworkformultiviewstereoimages
AT fanjiahao drimvsnetadepthresidualinferencenetworkformultiviewstereoimages