Cargando…
DRI-MVSNet: A depth residual inference network for multi-view stereo images
Three-dimensional (3D) image reconstruction is an important field of computer vision for restoring the 3D geometry of a given scene. Due to the demand for large amounts of memory, prevalent methods of 3D reconstruction yield inaccurate results, because of which the highly accuracy reconstruction of...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Public Library of Science
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8942269/ https://www.ncbi.nlm.nih.gov/pubmed/35320265 http://dx.doi.org/10.1371/journal.pone.0264721 |
_version_ | 1784673270298050560 |
---|---|
author | Li, Ying Li, Wenyue Zhao, Zhijie Fan, JiaHao |
author_facet | Li, Ying Li, Wenyue Zhao, Zhijie Fan, JiaHao |
author_sort | Li, Ying |
collection | PubMed |
description | Three-dimensional (3D) image reconstruction is an important field of computer vision for restoring the 3D geometry of a given scene. Due to the demand for large amounts of memory, prevalent methods of 3D reconstruction yield inaccurate results, because of which the highly accuracy reconstruction of a scene remains an outstanding challenge. This study proposes a cascaded depth residual inference network, called DRI-MVSNet, that uses a cross-view similarity-based feature map fusion module for residual inference. It involves three improvements. First, a combined module is used for processing channel-related and spatial information to capture the relevant contextual information and improve feature representation. It combines the channel attention mechanism and spatial pooling networks. Second, a cross-view similarity-based feature map fusion module is proposed that learns the similarity between pairs of pixel in each source and reference image at planes of different depths along the frustum of the reference camera. Third, a deep, multi-stage residual prediction module is designed to generate a high-precision depth map that uses a non-uniform depth sampling strategy to construct hypothetical depth planes. The results of extensive experiments show that DRI-MVSNet delivers competitive performance on the DTU and the Tanks & Temples datasets, and the accuracy and completeness of the point cloud reconstructed by it are significantly superior to those of state-of-the-art benchmarks. |
format | Online Article Text |
id | pubmed-8942269 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Public Library of Science |
record_format | MEDLINE/PubMed |
spelling | pubmed-89422692022-03-24 DRI-MVSNet: A depth residual inference network for multi-view stereo images Li, Ying Li, Wenyue Zhao, Zhijie Fan, JiaHao PLoS One Research Article Three-dimensional (3D) image reconstruction is an important field of computer vision for restoring the 3D geometry of a given scene. Due to the demand for large amounts of memory, prevalent methods of 3D reconstruction yield inaccurate results, because of which the highly accuracy reconstruction of a scene remains an outstanding challenge. This study proposes a cascaded depth residual inference network, called DRI-MVSNet, that uses a cross-view similarity-based feature map fusion module for residual inference. It involves three improvements. First, a combined module is used for processing channel-related and spatial information to capture the relevant contextual information and improve feature representation. It combines the channel attention mechanism and spatial pooling networks. Second, a cross-view similarity-based feature map fusion module is proposed that learns the similarity between pairs of pixel in each source and reference image at planes of different depths along the frustum of the reference camera. Third, a deep, multi-stage residual prediction module is designed to generate a high-precision depth map that uses a non-uniform depth sampling strategy to construct hypothetical depth planes. The results of extensive experiments show that DRI-MVSNet delivers competitive performance on the DTU and the Tanks & Temples datasets, and the accuracy and completeness of the point cloud reconstructed by it are significantly superior to those of state-of-the-art benchmarks. Public Library of Science 2022-03-23 /pmc/articles/PMC8942269/ /pubmed/35320265 http://dx.doi.org/10.1371/journal.pone.0264721 Text en © 2022 Li et al https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. |
spellingShingle | Research Article Li, Ying Li, Wenyue Zhao, Zhijie Fan, JiaHao DRI-MVSNet: A depth residual inference network for multi-view stereo images |
title | DRI-MVSNet: A depth residual inference network for multi-view stereo images |
title_full | DRI-MVSNet: A depth residual inference network for multi-view stereo images |
title_fullStr | DRI-MVSNet: A depth residual inference network for multi-view stereo images |
title_full_unstemmed | DRI-MVSNet: A depth residual inference network for multi-view stereo images |
title_short | DRI-MVSNet: A depth residual inference network for multi-view stereo images |
title_sort | dri-mvsnet: a depth residual inference network for multi-view stereo images |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8942269/ https://www.ncbi.nlm.nih.gov/pubmed/35320265 http://dx.doi.org/10.1371/journal.pone.0264721 |
work_keys_str_mv | AT liying drimvsnetadepthresidualinferencenetworkformultiviewstereoimages AT liwenyue drimvsnetadepthresidualinferencenetworkformultiviewstereoimages AT zhaozhijie drimvsnetadepthresidualinferencenetworkformultiviewstereoimages AT fanjiahao drimvsnetadepthresidualinferencenetworkformultiviewstereoimages |