Cargando…

Resolving Position Ambiguity of IMU-Based Human Pose with a Single RGB Camera

Human motion capture (MoCap) plays a key role in healthcare and human–robot collaboration. Some researchers have combined orientation measurements from inertial measurement units (IMUs) and positional inference from cameras to reconstruct the 3D human motion. Their works utilize multiple cameras or...

Descripción completa

Detalles Bibliográficos
Autores principales: Kaichi, Tomoya, Maruyama, Tsubasa, Tada, Mitsunori, Saito, Hideo
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7582626/
https://www.ncbi.nlm.nih.gov/pubmed/32977436
http://dx.doi.org/10.3390/s20195453
_version_ 1783599234788884480
author Kaichi, Tomoya
Maruyama, Tsubasa
Tada, Mitsunori
Saito, Hideo
author_facet Kaichi, Tomoya
Maruyama, Tsubasa
Tada, Mitsunori
Saito, Hideo
author_sort Kaichi, Tomoya
collection PubMed
description Human motion capture (MoCap) plays a key role in healthcare and human–robot collaboration. Some researchers have combined orientation measurements from inertial measurement units (IMUs) and positional inference from cameras to reconstruct the 3D human motion. Their works utilize multiple cameras or depth sensors to localize the human in three dimensions. Such multiple cameras are not always available in our daily life, but just a single camera attached in a smart IP devices has recently been popular. Therefore, we present a 3D pose estimation approach from IMUs and a single camera. In order to resolve the depth ambiguity of the single camera configuration and localize the global position of the subject, we present a constraint which optimizes the foot-ground contact points. The timing and 3D positions of the ground contact are calculated from the acceleration of IMUs on foot and geometric transformation of foot position detected on image, respectively. Since the results of pose estimation is greatly affected by the failure of the detection, we design the image-based constraints to handle the outliers of positional estimates. We evaluated the performance of our approach on public 3D human pose dataset. The experiments demonstrated that the proposed constraints contributed to improve the accuracy of pose estimation in single and multiple camera setting.
format Online
Article
Text
id pubmed-7582626
institution National Center for Biotechnology Information
language English
publishDate 2020
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-75826262020-10-28 Resolving Position Ambiguity of IMU-Based Human Pose with a Single RGB Camera Kaichi, Tomoya Maruyama, Tsubasa Tada, Mitsunori Saito, Hideo Sensors (Basel) Letter Human motion capture (MoCap) plays a key role in healthcare and human–robot collaboration. Some researchers have combined orientation measurements from inertial measurement units (IMUs) and positional inference from cameras to reconstruct the 3D human motion. Their works utilize multiple cameras or depth sensors to localize the human in three dimensions. Such multiple cameras are not always available in our daily life, but just a single camera attached in a smart IP devices has recently been popular. Therefore, we present a 3D pose estimation approach from IMUs and a single camera. In order to resolve the depth ambiguity of the single camera configuration and localize the global position of the subject, we present a constraint which optimizes the foot-ground contact points. The timing and 3D positions of the ground contact are calculated from the acceleration of IMUs on foot and geometric transformation of foot position detected on image, respectively. Since the results of pose estimation is greatly affected by the failure of the detection, we design the image-based constraints to handle the outliers of positional estimates. We evaluated the performance of our approach on public 3D human pose dataset. The experiments demonstrated that the proposed constraints contributed to improve the accuracy of pose estimation in single and multiple camera setting. MDPI 2020-09-23 /pmc/articles/PMC7582626/ /pubmed/32977436 http://dx.doi.org/10.3390/s20195453 Text en © 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Letter
Kaichi, Tomoya
Maruyama, Tsubasa
Tada, Mitsunori
Saito, Hideo
Resolving Position Ambiguity of IMU-Based Human Pose with a Single RGB Camera
title Resolving Position Ambiguity of IMU-Based Human Pose with a Single RGB Camera
title_full Resolving Position Ambiguity of IMU-Based Human Pose with a Single RGB Camera
title_fullStr Resolving Position Ambiguity of IMU-Based Human Pose with a Single RGB Camera
title_full_unstemmed Resolving Position Ambiguity of IMU-Based Human Pose with a Single RGB Camera
title_short Resolving Position Ambiguity of IMU-Based Human Pose with a Single RGB Camera
title_sort resolving position ambiguity of imu-based human pose with a single rgb camera
topic Letter
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7582626/
https://www.ncbi.nlm.nih.gov/pubmed/32977436
http://dx.doi.org/10.3390/s20195453
work_keys_str_mv AT kaichitomoya resolvingpositionambiguityofimubasedhumanposewithasinglergbcamera
AT maruyamatsubasa resolvingpositionambiguityofimubasedhumanposewithasinglergbcamera
AT tadamitsunori resolvingpositionambiguityofimubasedhumanposewithasinglergbcamera
AT saitohideo resolvingpositionambiguityofimubasedhumanposewithasinglergbcamera