Cargando…

DeepMoCap: Deep Optical Motion Capture Using Multiple Depth Sensors and Retro-Reflectors

In this paper, a marker-based, single-person optical motion capture method (DeepMoCap) is proposed using multiple spatio-temporally aligned infrared-depth sensors and retro-reflective straps and patches (reflectors). DeepMoCap explores motion capture by automatically localizing and labeling reflecto...

Descripción completa

Detalles Bibliográficos
Autores principales: Chatzitofis, Anargyros, Zarpalas, Dimitrios, Kollias, Stefanos, Daras, Petros
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2019
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6359336/
https://www.ncbi.nlm.nih.gov/pubmed/30642017
http://dx.doi.org/10.3390/s19020282
_version_ 1783392219416231936
author Chatzitofis, Anargyros
Zarpalas, Dimitrios
Kollias, Stefanos
Daras, Petros
author_facet Chatzitofis, Anargyros
Zarpalas, Dimitrios
Kollias, Stefanos
Daras, Petros
author_sort Chatzitofis, Anargyros
collection PubMed
description In this paper, a marker-based, single-person optical motion capture method (DeepMoCap) is proposed using multiple spatio-temporally aligned infrared-depth sensors and retro-reflective straps and patches (reflectors). DeepMoCap explores motion capture by automatically localizing and labeling reflectors on depth images and, subsequently, on 3D space. Introducing a non-parametric representation to encode the temporal correlation among pairs of colorized depthmaps and 3D optical flow frames, a multi-stage Fully Convolutional Network (FCN) architecture is proposed to jointly learn reflector locations and their temporal dependency among sequential frames. The extracted reflector 2D locations are spatially mapped in 3D space, resulting in robust 3D optical data extraction. The subject’s motion is efficiently captured by applying a template-based fitting technique on the extracted optical data. Two datasets have been created and made publicly available for evaluation purposes; one comprising multi-view depth and 3D optical flow annotated images (DMC2.5D), and a second, consisting of spatio-temporally aligned multi-view depth images along with skeleton, inertial and ground truth MoCap data (DMC3D). The FCN model outperforms its competitors on the DMC2.5D dataset using 2D Percentage of Correct Keypoints (PCK) metric, while the motion capture outcome is evaluated against RGB-D and inertial data fusion approaches on DMC3D, outperforming the next best method by [Formula: see text] in total 3D PCK accuracy.
format Online
Article
Text
id pubmed-6359336
institution National Center for Biotechnology Information
language English
publishDate 2019
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-63593362019-02-06 DeepMoCap: Deep Optical Motion Capture Using Multiple Depth Sensors and Retro-Reflectors Chatzitofis, Anargyros Zarpalas, Dimitrios Kollias, Stefanos Daras, Petros Sensors (Basel) Article In this paper, a marker-based, single-person optical motion capture method (DeepMoCap) is proposed using multiple spatio-temporally aligned infrared-depth sensors and retro-reflective straps and patches (reflectors). DeepMoCap explores motion capture by automatically localizing and labeling reflectors on depth images and, subsequently, on 3D space. Introducing a non-parametric representation to encode the temporal correlation among pairs of colorized depthmaps and 3D optical flow frames, a multi-stage Fully Convolutional Network (FCN) architecture is proposed to jointly learn reflector locations and their temporal dependency among sequential frames. The extracted reflector 2D locations are spatially mapped in 3D space, resulting in robust 3D optical data extraction. The subject’s motion is efficiently captured by applying a template-based fitting technique on the extracted optical data. Two datasets have been created and made publicly available for evaluation purposes; one comprising multi-view depth and 3D optical flow annotated images (DMC2.5D), and a second, consisting of spatio-temporally aligned multi-view depth images along with skeleton, inertial and ground truth MoCap data (DMC3D). The FCN model outperforms its competitors on the DMC2.5D dataset using 2D Percentage of Correct Keypoints (PCK) metric, while the motion capture outcome is evaluated against RGB-D and inertial data fusion approaches on DMC3D, outperforming the next best method by [Formula: see text] in total 3D PCK accuracy. MDPI 2019-01-11 /pmc/articles/PMC6359336/ /pubmed/30642017 http://dx.doi.org/10.3390/s19020282 Text en © 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Chatzitofis, Anargyros
Zarpalas, Dimitrios
Kollias, Stefanos
Daras, Petros
DeepMoCap: Deep Optical Motion Capture Using Multiple Depth Sensors and Retro-Reflectors
title DeepMoCap: Deep Optical Motion Capture Using Multiple Depth Sensors and Retro-Reflectors
title_full DeepMoCap: Deep Optical Motion Capture Using Multiple Depth Sensors and Retro-Reflectors
title_fullStr DeepMoCap: Deep Optical Motion Capture Using Multiple Depth Sensors and Retro-Reflectors
title_full_unstemmed DeepMoCap: Deep Optical Motion Capture Using Multiple Depth Sensors and Retro-Reflectors
title_short DeepMoCap: Deep Optical Motion Capture Using Multiple Depth Sensors and Retro-Reflectors
title_sort deepmocap: deep optical motion capture using multiple depth sensors and retro-reflectors
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6359336/
https://www.ncbi.nlm.nih.gov/pubmed/30642017
http://dx.doi.org/10.3390/s19020282
work_keys_str_mv AT chatzitofisanargyros deepmocapdeepopticalmotioncaptureusingmultipledepthsensorsandretroreflectors
AT zarpalasdimitrios deepmocapdeepopticalmotioncaptureusingmultipledepthsensorsandretroreflectors
AT kolliasstefanos deepmocapdeepopticalmotioncaptureusingmultipledepthsensorsandretroreflectors
AT daraspetros deepmocapdeepopticalmotioncaptureusingmultipledepthsensorsandretroreflectors