Cargando…

Attention-Based 3D Human Pose Sequence Refinement Network

Three-dimensional human mesh reconstruction from a single video has made much progress in recent years due to the advances in deep learning. However, previous methods still often reconstruct temporally noisy pose and mesh sequences given in-the-wild video data. To address this problem, we propose a...

Descripción completa

Detalles Bibliográficos
Autores principales: Kim, Do-Yeop, Chang, Ju-Yong
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8271996/
https://www.ncbi.nlm.nih.gov/pubmed/34283128
http://dx.doi.org/10.3390/s21134572
_version_ 1783721122181677056
author Kim, Do-Yeop
Chang, Ju-Yong
author_facet Kim, Do-Yeop
Chang, Ju-Yong
author_sort Kim, Do-Yeop
collection PubMed
description Three-dimensional human mesh reconstruction from a single video has made much progress in recent years due to the advances in deep learning. However, previous methods still often reconstruct temporally noisy pose and mesh sequences given in-the-wild video data. To address this problem, we propose a human pose refinement network (HPR-Net) based on a non-local attention mechanism. The pipeline of the proposed framework consists of a weight-regression module, a weighted-averaging module, and a skinned multi-person linear (SMPL) module. First, the weight-regression module creates pose affinity weights from a 3D human pose sequence represented in a unit quaternion form. Next, the weighted-averaging module generates a refined 3D pose sequence by performing temporal weighted averaging using the generated affinity weights. Finally, the refined pose sequence is converted into a human mesh sequence using the SMPL module. HPR-Net is a simple but effective post-processing network that can substantially improve the accuracy and temporal smoothness of 3D human mesh sequences obtained from an input video by existing human mesh reconstruction methods. Our experiments show that the noisy results of the existing methods are consistently improved using the proposed method on various real datasets. Notably, our proposed method reduces the pose and acceleration errors of VIBE, the existing state-of-the-art human mesh reconstruction method, by 1.4% and 66.5%, respectively, on the 3DPW dataset.
format Online
Article
Text
id pubmed-8271996
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-82719962021-07-11 Attention-Based 3D Human Pose Sequence Refinement Network Kim, Do-Yeop Chang, Ju-Yong Sensors (Basel) Article Three-dimensional human mesh reconstruction from a single video has made much progress in recent years due to the advances in deep learning. However, previous methods still often reconstruct temporally noisy pose and mesh sequences given in-the-wild video data. To address this problem, we propose a human pose refinement network (HPR-Net) based on a non-local attention mechanism. The pipeline of the proposed framework consists of a weight-regression module, a weighted-averaging module, and a skinned multi-person linear (SMPL) module. First, the weight-regression module creates pose affinity weights from a 3D human pose sequence represented in a unit quaternion form. Next, the weighted-averaging module generates a refined 3D pose sequence by performing temporal weighted averaging using the generated affinity weights. Finally, the refined pose sequence is converted into a human mesh sequence using the SMPL module. HPR-Net is a simple but effective post-processing network that can substantially improve the accuracy and temporal smoothness of 3D human mesh sequences obtained from an input video by existing human mesh reconstruction methods. Our experiments show that the noisy results of the existing methods are consistently improved using the proposed method on various real datasets. Notably, our proposed method reduces the pose and acceleration errors of VIBE, the existing state-of-the-art human mesh reconstruction method, by 1.4% and 66.5%, respectively, on the 3DPW dataset. MDPI 2021-07-03 /pmc/articles/PMC8271996/ /pubmed/34283128 http://dx.doi.org/10.3390/s21134572 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Kim, Do-Yeop
Chang, Ju-Yong
Attention-Based 3D Human Pose Sequence Refinement Network
title Attention-Based 3D Human Pose Sequence Refinement Network
title_full Attention-Based 3D Human Pose Sequence Refinement Network
title_fullStr Attention-Based 3D Human Pose Sequence Refinement Network
title_full_unstemmed Attention-Based 3D Human Pose Sequence Refinement Network
title_short Attention-Based 3D Human Pose Sequence Refinement Network
title_sort attention-based 3d human pose sequence refinement network
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8271996/
https://www.ncbi.nlm.nih.gov/pubmed/34283128
http://dx.doi.org/10.3390/s21134572
work_keys_str_mv AT kimdoyeop attentionbased3dhumanposesequencerefinementnetwork
AT changjuyong attentionbased3dhumanposesequencerefinementnetwork