Cargando…

Fusion of Multiple Lidars and Inertial Sensors for the Real-Time Pose Tracking of Human Motion

Today, enhancement in sensing technology enables the use of multiple sensors to track human motion/activity precisely. Tracking human motion has various applications, such as fitness training, healthcare, rehabilitation, human-computer interaction, virtual reality, and activity recognition. Therefor...

Descripción completa

Detalles Bibliográficos
Autores principales: Patil, Ashok Kumar, Balasubramanyam, Adithya, Ryu, Jae Yeong, B N, Pavan Kumar, Chakravarthi, Bharatesh, Chai, Young Ho
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7570691/
https://www.ncbi.nlm.nih.gov/pubmed/32961918
http://dx.doi.org/10.3390/s20185342
_version_ 1783597005512114176
author Patil, Ashok Kumar
Balasubramanyam, Adithya
Ryu, Jae Yeong
B N, Pavan Kumar
Chakravarthi, Bharatesh
Chai, Young Ho
author_facet Patil, Ashok Kumar
Balasubramanyam, Adithya
Ryu, Jae Yeong
B N, Pavan Kumar
Chakravarthi, Bharatesh
Chai, Young Ho
author_sort Patil, Ashok Kumar
collection PubMed
description Today, enhancement in sensing technology enables the use of multiple sensors to track human motion/activity precisely. Tracking human motion has various applications, such as fitness training, healthcare, rehabilitation, human-computer interaction, virtual reality, and activity recognition. Therefore, the fusion of multiple sensors creates new opportunities to develop and improve an existing system. This paper proposes a pose-tracking system by fusing multiple three-dimensional (3D) light detection and ranging (lidar) and inertial measurement unit (IMU) sensors. The initial step estimates the human skeletal parameters proportional to the target user’s height by extracting the point cloud from lidars. Next, IMUs are used to capture the orientation of each skeleton segment and estimate the respective joint positions. In the final stage, the displacement drift in the position is corrected by fusing the data from both sensors in real time. The installation setup is relatively effortless, flexible for sensor locations, and delivers results comparable to the state-of-the-art pose-tracking system. We evaluated the proposed system regarding its accuracy in the user’s height estimation, full-body joint position estimation, and reconstruction of the 3D avatar. We used a publicly available dataset for the experimental evaluation wherever possible. The results reveal that the accuracy of height and the position estimation is well within an acceptable range of ±3–5 cm. The reconstruction of the motion based on the publicly available dataset and our data is precise and realistic.
format Online
Article
Text
id pubmed-7570691
institution National Center for Biotechnology Information
language English
publishDate 2020
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-75706912020-10-28 Fusion of Multiple Lidars and Inertial Sensors for the Real-Time Pose Tracking of Human Motion Patil, Ashok Kumar Balasubramanyam, Adithya Ryu, Jae Yeong B N, Pavan Kumar Chakravarthi, Bharatesh Chai, Young Ho Sensors (Basel) Article Today, enhancement in sensing technology enables the use of multiple sensors to track human motion/activity precisely. Tracking human motion has various applications, such as fitness training, healthcare, rehabilitation, human-computer interaction, virtual reality, and activity recognition. Therefore, the fusion of multiple sensors creates new opportunities to develop and improve an existing system. This paper proposes a pose-tracking system by fusing multiple three-dimensional (3D) light detection and ranging (lidar) and inertial measurement unit (IMU) sensors. The initial step estimates the human skeletal parameters proportional to the target user’s height by extracting the point cloud from lidars. Next, IMUs are used to capture the orientation of each skeleton segment and estimate the respective joint positions. In the final stage, the displacement drift in the position is corrected by fusing the data from both sensors in real time. The installation setup is relatively effortless, flexible for sensor locations, and delivers results comparable to the state-of-the-art pose-tracking system. We evaluated the proposed system regarding its accuracy in the user’s height estimation, full-body joint position estimation, and reconstruction of the 3D avatar. We used a publicly available dataset for the experimental evaluation wherever possible. The results reveal that the accuracy of height and the position estimation is well within an acceptable range of ±3–5 cm. The reconstruction of the motion based on the publicly available dataset and our data is precise and realistic. MDPI 2020-09-18 /pmc/articles/PMC7570691/ /pubmed/32961918 http://dx.doi.org/10.3390/s20185342 Text en © 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Patil, Ashok Kumar
Balasubramanyam, Adithya
Ryu, Jae Yeong
B N, Pavan Kumar
Chakravarthi, Bharatesh
Chai, Young Ho
Fusion of Multiple Lidars and Inertial Sensors for the Real-Time Pose Tracking of Human Motion
title Fusion of Multiple Lidars and Inertial Sensors for the Real-Time Pose Tracking of Human Motion
title_full Fusion of Multiple Lidars and Inertial Sensors for the Real-Time Pose Tracking of Human Motion
title_fullStr Fusion of Multiple Lidars and Inertial Sensors for the Real-Time Pose Tracking of Human Motion
title_full_unstemmed Fusion of Multiple Lidars and Inertial Sensors for the Real-Time Pose Tracking of Human Motion
title_short Fusion of Multiple Lidars and Inertial Sensors for the Real-Time Pose Tracking of Human Motion
title_sort fusion of multiple lidars and inertial sensors for the real-time pose tracking of human motion
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7570691/
https://www.ncbi.nlm.nih.gov/pubmed/32961918
http://dx.doi.org/10.3390/s20185342
work_keys_str_mv AT patilashokkumar fusionofmultiplelidarsandinertialsensorsfortherealtimeposetrackingofhumanmotion
AT balasubramanyamadithya fusionofmultiplelidarsandinertialsensorsfortherealtimeposetrackingofhumanmotion
AT ryujaeyeong fusionofmultiplelidarsandinertialsensorsfortherealtimeposetrackingofhumanmotion
AT bnpavankumar fusionofmultiplelidarsandinertialsensorsfortherealtimeposetrackingofhumanmotion
AT chakravarthibharatesh fusionofmultiplelidarsandinertialsensorsfortherealtimeposetrackingofhumanmotion
AT chaiyoungho fusionofmultiplelidarsandinertialsensorsfortherealtimeposetrackingofhumanmotion