Cargando…
3D real-time human reconstruction with a single RGBD camera
3D human reconstruction is an important technology connecting the real world and the virtual world, but most of previous work needs expensive computing resources, making it difficult in real-time scenarios. We propose a lightweight human body reconstruction system based on parametric model, which em...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Springer US
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9343569/ https://www.ncbi.nlm.nih.gov/pubmed/35937202 http://dx.doi.org/10.1007/s10489-022-03969-4 |
_version_ | 1784761018507853824 |
---|---|
author | Lu, Yang Yu, Han Ni, Wei Song, Liang |
author_facet | Lu, Yang Yu, Han Ni, Wei Song, Liang |
author_sort | Lu, Yang |
collection | PubMed |
description | 3D human reconstruction is an important technology connecting the real world and the virtual world, but most of previous work needs expensive computing resources, making it difficult in real-time scenarios. We propose a lightweight human body reconstruction system based on parametric model, which employs only one RGBD camera as input. To generate a human model end to end, we build a fast and lightweight deep-learning network named Fast Body Net (FBN). The network pays more attention on the face and hands to enrich the local details. Additionally, we train a denoising auto-encoder to reduce unreasonable states of human model. Due to the lack of human dataset based on RGBD images, we propose an Indoor-Human dataset to train the network, which contains a total of 2500 frames of action data of five actors collected by Azure Kinect camera. Depth images avoid using RGB to extract depth features, which makes FBN lightweight and high-speed in reconstructing parametric human model. Qualitative and quantitative analysis on experimental results show that our method can improve at least 57% in efficiency with similar accuracy, as compared to state-of-the-art methods. Through our study, it is also demonstrated that consumer-grade RGBD cameras can provide great applications in real-time display and interaction for virtual reality. |
format | Online Article Text |
id | pubmed-9343569 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Springer US |
record_format | MEDLINE/PubMed |
spelling | pubmed-93435692022-08-02 3D real-time human reconstruction with a single RGBD camera Lu, Yang Yu, Han Ni, Wei Song, Liang Appl Intell (Dordr) Article 3D human reconstruction is an important technology connecting the real world and the virtual world, but most of previous work needs expensive computing resources, making it difficult in real-time scenarios. We propose a lightweight human body reconstruction system based on parametric model, which employs only one RGBD camera as input. To generate a human model end to end, we build a fast and lightweight deep-learning network named Fast Body Net (FBN). The network pays more attention on the face and hands to enrich the local details. Additionally, we train a denoising auto-encoder to reduce unreasonable states of human model. Due to the lack of human dataset based on RGBD images, we propose an Indoor-Human dataset to train the network, which contains a total of 2500 frames of action data of five actors collected by Azure Kinect camera. Depth images avoid using RGB to extract depth features, which makes FBN lightweight and high-speed in reconstructing parametric human model. Qualitative and quantitative analysis on experimental results show that our method can improve at least 57% in efficiency with similar accuracy, as compared to state-of-the-art methods. Through our study, it is also demonstrated that consumer-grade RGBD cameras can provide great applications in real-time display and interaction for virtual reality. Springer US 2022-08-02 2023 /pmc/articles/PMC9343569/ /pubmed/35937202 http://dx.doi.org/10.1007/s10489-022-03969-4 Text en © The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022 This article is made available via the PMC Open Access Subset for unrestricted research re-use and secondary analysis in any form or by any means with acknowledgement of the original source. These permissions are granted for the duration of the World Health Organization (WHO) declaration of COVID-19 as a global pandemic. |
spellingShingle | Article Lu, Yang Yu, Han Ni, Wei Song, Liang 3D real-time human reconstruction with a single RGBD camera |
title | 3D real-time human reconstruction with a single RGBD camera |
title_full | 3D real-time human reconstruction with a single RGBD camera |
title_fullStr | 3D real-time human reconstruction with a single RGBD camera |
title_full_unstemmed | 3D real-time human reconstruction with a single RGBD camera |
title_short | 3D real-time human reconstruction with a single RGBD camera |
title_sort | 3d real-time human reconstruction with a single rgbd camera |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9343569/ https://www.ncbi.nlm.nih.gov/pubmed/35937202 http://dx.doi.org/10.1007/s10489-022-03969-4 |
work_keys_str_mv | AT luyang 3drealtimehumanreconstructionwithasinglergbdcamera AT yuhan 3drealtimehumanreconstructionwithasinglergbdcamera AT niwei 3drealtimehumanreconstructionwithasinglergbdcamera AT songliang 3drealtimehumanreconstructionwithasinglergbdcamera |