Cargando…

Real-Time 3D Reconstruction Method Based on Monocular Vision

Real-time 3D reconstruction is one of the current popular research directions of computer vision, and it has become the core technology in the fields of virtual reality, industrialized automatic systems, and mobile robot path planning. Currently, there are three main problems in the real-time 3D rec...

Descripción completa

Detalles Bibliográficos
Autores principales: Jia, Qingyu, Chang, Liang, Qiang, Baohua, Zhang, Shihao, Xie, Wu, Yang, Xianyi, Sun, Yangchang, Yang, Minghao
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8434368/
https://www.ncbi.nlm.nih.gov/pubmed/34502800
http://dx.doi.org/10.3390/s21175909
_version_ 1783751582201937920
author Jia, Qingyu
Chang, Liang
Qiang, Baohua
Zhang, Shihao
Xie, Wu
Yang, Xianyi
Sun, Yangchang
Yang, Minghao
author_facet Jia, Qingyu
Chang, Liang
Qiang, Baohua
Zhang, Shihao
Xie, Wu
Yang, Xianyi
Sun, Yangchang
Yang, Minghao
author_sort Jia, Qingyu
collection PubMed
description Real-time 3D reconstruction is one of the current popular research directions of computer vision, and it has become the core technology in the fields of virtual reality, industrialized automatic systems, and mobile robot path planning. Currently, there are three main problems in the real-time 3D reconstruction field. Firstly, it is expensive. It requires more varied sensors, so it is less convenient. Secondly, the reconstruction speed is slow, and the 3D model cannot be established accurately in real time. Thirdly, the reconstruction error is large, which cannot meet the requirements of scenes with accuracy. For this reason, we propose a real-time 3D reconstruction method based on monocular vision in this paper. Firstly, a single RGB-D camera is used to collect visual information in real time, and the YOLACT++ network is used to identify and segment the visual information to extract part of the important visual information. Secondly, we combine the three stages of depth recovery, depth optimization, and deep fusion to propose a three-dimensional position estimation method based on deep learning for joint coding of visual information. It can reduce the depth error caused by the depth measurement process, and the accurate 3D point values of the segmented image can be obtained directly. Finally, we propose a method based on the limited outlier adjustment of the cluster center distance to optimize the three-dimensional point values obtained above. It improves the real-time reconstruction accuracy and obtains the three-dimensional model of the object in real time. Experimental results show that this method only needs a single RGB-D camera, which is not only low cost and convenient to use, but also significantly improves the speed and accuracy of 3D reconstruction.
format Online
Article
Text
id pubmed-8434368
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-84343682021-09-12 Real-Time 3D Reconstruction Method Based on Monocular Vision Jia, Qingyu Chang, Liang Qiang, Baohua Zhang, Shihao Xie, Wu Yang, Xianyi Sun, Yangchang Yang, Minghao Sensors (Basel) Article Real-time 3D reconstruction is one of the current popular research directions of computer vision, and it has become the core technology in the fields of virtual reality, industrialized automatic systems, and mobile robot path planning. Currently, there are three main problems in the real-time 3D reconstruction field. Firstly, it is expensive. It requires more varied sensors, so it is less convenient. Secondly, the reconstruction speed is slow, and the 3D model cannot be established accurately in real time. Thirdly, the reconstruction error is large, which cannot meet the requirements of scenes with accuracy. For this reason, we propose a real-time 3D reconstruction method based on monocular vision in this paper. Firstly, a single RGB-D camera is used to collect visual information in real time, and the YOLACT++ network is used to identify and segment the visual information to extract part of the important visual information. Secondly, we combine the three stages of depth recovery, depth optimization, and deep fusion to propose a three-dimensional position estimation method based on deep learning for joint coding of visual information. It can reduce the depth error caused by the depth measurement process, and the accurate 3D point values of the segmented image can be obtained directly. Finally, we propose a method based on the limited outlier adjustment of the cluster center distance to optimize the three-dimensional point values obtained above. It improves the real-time reconstruction accuracy and obtains the three-dimensional model of the object in real time. Experimental results show that this method only needs a single RGB-D camera, which is not only low cost and convenient to use, but also significantly improves the speed and accuracy of 3D reconstruction. MDPI 2021-09-02 /pmc/articles/PMC8434368/ /pubmed/34502800 http://dx.doi.org/10.3390/s21175909 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Jia, Qingyu
Chang, Liang
Qiang, Baohua
Zhang, Shihao
Xie, Wu
Yang, Xianyi
Sun, Yangchang
Yang, Minghao
Real-Time 3D Reconstruction Method Based on Monocular Vision
title Real-Time 3D Reconstruction Method Based on Monocular Vision
title_full Real-Time 3D Reconstruction Method Based on Monocular Vision
title_fullStr Real-Time 3D Reconstruction Method Based on Monocular Vision
title_full_unstemmed Real-Time 3D Reconstruction Method Based on Monocular Vision
title_short Real-Time 3D Reconstruction Method Based on Monocular Vision
title_sort real-time 3d reconstruction method based on monocular vision
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8434368/
https://www.ncbi.nlm.nih.gov/pubmed/34502800
http://dx.doi.org/10.3390/s21175909
work_keys_str_mv AT jiaqingyu realtime3dreconstructionmethodbasedonmonocularvision
AT changliang realtime3dreconstructionmethodbasedonmonocularvision
AT qiangbaohua realtime3dreconstructionmethodbasedonmonocularvision
AT zhangshihao realtime3dreconstructionmethodbasedonmonocularvision
AT xiewu realtime3dreconstructionmethodbasedonmonocularvision
AT yangxianyi realtime3dreconstructionmethodbasedonmonocularvision
AT sunyangchang realtime3dreconstructionmethodbasedonmonocularvision
AT yangminghao realtime3dreconstructionmethodbasedonmonocularvision