Cargando…

Center point to pose: Multiple views 3D human pose estimation for multi-person

3D human pose estimation has always been an important task in computer vision, especially in crowded scenes where multiple people interact with each other. There are many state-of-the-arts for object detection based on single view. However, recovering the location of people is complicated in crowded...

Descripción completa

Detalles Bibliográficos
Autores principales: Liu, Huan, Wu, Jian, He, Rui
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Public Library of Science 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9469997/
https://www.ncbi.nlm.nih.gov/pubmed/36099276
http://dx.doi.org/10.1371/journal.pone.0274450
_version_ 1784788755044892672
author Liu, Huan
Wu, Jian
He, Rui
author_facet Liu, Huan
Wu, Jian
He, Rui
author_sort Liu, Huan
collection PubMed
description 3D human pose estimation has always been an important task in computer vision, especially in crowded scenes where multiple people interact with each other. There are many state-of-the-arts for object detection based on single view. However, recovering the location of people is complicated in crowded and occluded scenes due to the lack of depth information for single view, which is the lack of robustness. Multi-view Human Pose Estimation for Multi-Person became an effective approach. The previous multi-view 3D human pose estimation method can be attributed to a strategy to associate the joints of the same person from 2D pose estimation. However, the incompleteness and noise of the 2D pose are inevitable. In addition, how to associate the joints itself is challenging. To solve this issue, we propose a CTP (Center Point to Pose) network based on multi-view which directly operates in the 3D space. The 2D joint features in all cameras are projected into 3D voxel space. Our CTP network regresses the center of one person as the location, and the 3D bounding box as the activity area of one person. Then our CTP network estimates detailed 3D pose for each bounding box. Besides, our CTP network is Non-Maximum Suppression free at the stage of regressing the center of one person, which makes it more efficient and simpler. Our method outperforms competitively on several public datasets which shows the efficacy of our center point to pose network representation.
format Online
Article
Text
id pubmed-9469997
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-94699972022-09-14 Center point to pose: Multiple views 3D human pose estimation for multi-person Liu, Huan Wu, Jian He, Rui PLoS One Research Article 3D human pose estimation has always been an important task in computer vision, especially in crowded scenes where multiple people interact with each other. There are many state-of-the-arts for object detection based on single view. However, recovering the location of people is complicated in crowded and occluded scenes due to the lack of depth information for single view, which is the lack of robustness. Multi-view Human Pose Estimation for Multi-Person became an effective approach. The previous multi-view 3D human pose estimation method can be attributed to a strategy to associate the joints of the same person from 2D pose estimation. However, the incompleteness and noise of the 2D pose are inevitable. In addition, how to associate the joints itself is challenging. To solve this issue, we propose a CTP (Center Point to Pose) network based on multi-view which directly operates in the 3D space. The 2D joint features in all cameras are projected into 3D voxel space. Our CTP network regresses the center of one person as the location, and the 3D bounding box as the activity area of one person. Then our CTP network estimates detailed 3D pose for each bounding box. Besides, our CTP network is Non-Maximum Suppression free at the stage of regressing the center of one person, which makes it more efficient and simpler. Our method outperforms competitively on several public datasets which shows the efficacy of our center point to pose network representation. Public Library of Science 2022-09-13 /pmc/articles/PMC9469997/ /pubmed/36099276 http://dx.doi.org/10.1371/journal.pone.0274450 Text en © 2022 Liu et al https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
spellingShingle Research Article
Liu, Huan
Wu, Jian
He, Rui
Center point to pose: Multiple views 3D human pose estimation for multi-person
title Center point to pose: Multiple views 3D human pose estimation for multi-person
title_full Center point to pose: Multiple views 3D human pose estimation for multi-person
title_fullStr Center point to pose: Multiple views 3D human pose estimation for multi-person
title_full_unstemmed Center point to pose: Multiple views 3D human pose estimation for multi-person
title_short Center point to pose: Multiple views 3D human pose estimation for multi-person
title_sort center point to pose: multiple views 3d human pose estimation for multi-person
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9469997/
https://www.ncbi.nlm.nih.gov/pubmed/36099276
http://dx.doi.org/10.1371/journal.pone.0274450
work_keys_str_mv AT liuhuan centerpointtoposemultipleviews3dhumanposeestimationformultiperson
AT wujian centerpointtoposemultipleviews3dhumanposeestimationformultiperson
AT herui centerpointtoposemultipleviews3dhumanposeestimationformultiperson