Cargando…
Fusion Poser: 3D Human Pose Estimation Using Sparse IMUs and Head Trackers in Real Time
The motion capture method using sparse inertial sensors is an approach for solving the occlusion and economic problems in vision-based methods, which is suitable for virtual reality applications and works in complex environments. However, VR applications need to track the location of the user in rea...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9269439/ https://www.ncbi.nlm.nih.gov/pubmed/35808342 http://dx.doi.org/10.3390/s22134846 |
_version_ | 1784744237069238272 |
---|---|
author | Kim, Meejin Lee, Sukwon |
author_facet | Kim, Meejin Lee, Sukwon |
author_sort | Kim, Meejin |
collection | PubMed |
description | The motion capture method using sparse inertial sensors is an approach for solving the occlusion and economic problems in vision-based methods, which is suitable for virtual reality applications and works in complex environments. However, VR applications need to track the location of the user in real-world space, which is hard to obtain using only inertial sensors. In this paper, we present Fusion Poser, which combines the deep learning-based pose estimation and location tracking method with six inertial measurement units and a head tracking sensor that provides head-mounted displays. To estimate human poses, we propose a bidirectional recurrent neural network with a convolutional long short-term memory layer that achieves higher accuracy and stability by preserving spatio-temporal properties. To locate a user with real-world coordinates, our method integrates the results of an estimated joint pose with the pose of the tracker. To train the model, we gathered public motion capture datasets of synthesized IMU measurement data, as well as creating a real-world dataset. In the evaluation, our method showed higher accuracy and a more robust estimation performance, especially when the user adopted lower poses, such as a squat or a bow. |
format | Online Article Text |
id | pubmed-9269439 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-92694392022-07-09 Fusion Poser: 3D Human Pose Estimation Using Sparse IMUs and Head Trackers in Real Time Kim, Meejin Lee, Sukwon Sensors (Basel) Article The motion capture method using sparse inertial sensors is an approach for solving the occlusion and economic problems in vision-based methods, which is suitable for virtual reality applications and works in complex environments. However, VR applications need to track the location of the user in real-world space, which is hard to obtain using only inertial sensors. In this paper, we present Fusion Poser, which combines the deep learning-based pose estimation and location tracking method with six inertial measurement units and a head tracking sensor that provides head-mounted displays. To estimate human poses, we propose a bidirectional recurrent neural network with a convolutional long short-term memory layer that achieves higher accuracy and stability by preserving spatio-temporal properties. To locate a user with real-world coordinates, our method integrates the results of an estimated joint pose with the pose of the tracker. To train the model, we gathered public motion capture datasets of synthesized IMU measurement data, as well as creating a real-world dataset. In the evaluation, our method showed higher accuracy and a more robust estimation performance, especially when the user adopted lower poses, such as a squat or a bow. MDPI 2022-06-27 /pmc/articles/PMC9269439/ /pubmed/35808342 http://dx.doi.org/10.3390/s22134846 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Kim, Meejin Lee, Sukwon Fusion Poser: 3D Human Pose Estimation Using Sparse IMUs and Head Trackers in Real Time |
title | Fusion Poser: 3D Human Pose Estimation Using Sparse IMUs and Head Trackers in Real Time |
title_full | Fusion Poser: 3D Human Pose Estimation Using Sparse IMUs and Head Trackers in Real Time |
title_fullStr | Fusion Poser: 3D Human Pose Estimation Using Sparse IMUs and Head Trackers in Real Time |
title_full_unstemmed | Fusion Poser: 3D Human Pose Estimation Using Sparse IMUs and Head Trackers in Real Time |
title_short | Fusion Poser: 3D Human Pose Estimation Using Sparse IMUs and Head Trackers in Real Time |
title_sort | fusion poser: 3d human pose estimation using sparse imus and head trackers in real time |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9269439/ https://www.ncbi.nlm.nih.gov/pubmed/35808342 http://dx.doi.org/10.3390/s22134846 |
work_keys_str_mv | AT kimmeejin fusionposer3dhumanposeestimationusingsparseimusandheadtrackersinrealtime AT leesukwon fusionposer3dhumanposeestimationusingsparseimusandheadtrackersinrealtime |