Cargando…

Adaptive Local Spatiotemporal Features from RGB-D Data for One-Shot Learning Gesture Recognition

Noise and constant empirical motion constraints affect the extraction of distinctive spatiotemporal features from one or a few samples per gesture class. To tackle these problems, an adaptive local spatiotemporal feature (ALSTF) using fused RGB-D data is proposed. First, motion regions of interest (...

Descripción completa

Detalles Bibliográficos
Autores principales: Lin, Jia, Ruan, Xiaogang, Yu, Naigong, Yang, Yee-Hong
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2016
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5191150/
https://www.ncbi.nlm.nih.gov/pubmed/27999337
http://dx.doi.org/10.3390/s16122171
_version_ 1782487567775563776
author Lin, Jia
Ruan, Xiaogang
Yu, Naigong
Yang, Yee-Hong
author_facet Lin, Jia
Ruan, Xiaogang
Yu, Naigong
Yang, Yee-Hong
author_sort Lin, Jia
collection PubMed
description Noise and constant empirical motion constraints affect the extraction of distinctive spatiotemporal features from one or a few samples per gesture class. To tackle these problems, an adaptive local spatiotemporal feature (ALSTF) using fused RGB-D data is proposed. First, motion regions of interest (MRoIs) are adaptively extracted using grayscale and depth velocity variance information to greatly reduce the impact of noise. Then, corners are used as keypoints if their depth, and velocities of grayscale and of depth meet several adaptive local constraints in each MRoI. With further filtering of noise, an accurate and sufficient number of keypoints is obtained within the desired moving body parts (MBPs). Finally, four kinds of multiple descriptors are calculated and combined in extended gradient and motion spaces to represent the appearance and motion features of gestures. The experimental results on the ChaLearn gesture, CAD-60 and MSRDailyActivity3D datasets demonstrate that the proposed feature achieves higher performance compared with published state-of-the-art approaches under the one-shot learning setting and comparable accuracy under the leave-one-out cross validation.
format Online
Article
Text
id pubmed-5191150
institution National Center for Biotechnology Information
language English
publishDate 2016
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-51911502017-01-03 Adaptive Local Spatiotemporal Features from RGB-D Data for One-Shot Learning Gesture Recognition Lin, Jia Ruan, Xiaogang Yu, Naigong Yang, Yee-Hong Sensors (Basel) Article Noise and constant empirical motion constraints affect the extraction of distinctive spatiotemporal features from one or a few samples per gesture class. To tackle these problems, an adaptive local spatiotemporal feature (ALSTF) using fused RGB-D data is proposed. First, motion regions of interest (MRoIs) are adaptively extracted using grayscale and depth velocity variance information to greatly reduce the impact of noise. Then, corners are used as keypoints if their depth, and velocities of grayscale and of depth meet several adaptive local constraints in each MRoI. With further filtering of noise, an accurate and sufficient number of keypoints is obtained within the desired moving body parts (MBPs). Finally, four kinds of multiple descriptors are calculated and combined in extended gradient and motion spaces to represent the appearance and motion features of gestures. The experimental results on the ChaLearn gesture, CAD-60 and MSRDailyActivity3D datasets demonstrate that the proposed feature achieves higher performance compared with published state-of-the-art approaches under the one-shot learning setting and comparable accuracy under the leave-one-out cross validation. MDPI 2016-12-17 /pmc/articles/PMC5191150/ /pubmed/27999337 http://dx.doi.org/10.3390/s16122171 Text en © 2016 by the authors; licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC-BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Lin, Jia
Ruan, Xiaogang
Yu, Naigong
Yang, Yee-Hong
Adaptive Local Spatiotemporal Features from RGB-D Data for One-Shot Learning Gesture Recognition
title Adaptive Local Spatiotemporal Features from RGB-D Data for One-Shot Learning Gesture Recognition
title_full Adaptive Local Spatiotemporal Features from RGB-D Data for One-Shot Learning Gesture Recognition
title_fullStr Adaptive Local Spatiotemporal Features from RGB-D Data for One-Shot Learning Gesture Recognition
title_full_unstemmed Adaptive Local Spatiotemporal Features from RGB-D Data for One-Shot Learning Gesture Recognition
title_short Adaptive Local Spatiotemporal Features from RGB-D Data for One-Shot Learning Gesture Recognition
title_sort adaptive local spatiotemporal features from rgb-d data for one-shot learning gesture recognition
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5191150/
https://www.ncbi.nlm.nih.gov/pubmed/27999337
http://dx.doi.org/10.3390/s16122171
work_keys_str_mv AT linjia adaptivelocalspatiotemporalfeaturesfromrgbddataforoneshotlearninggesturerecognition
AT ruanxiaogang adaptivelocalspatiotemporalfeaturesfromrgbddataforoneshotlearninggesturerecognition
AT yunaigong adaptivelocalspatiotemporalfeaturesfromrgbddataforoneshotlearninggesturerecognition
AT yangyeehong adaptivelocalspatiotemporalfeaturesfromrgbddataforoneshotlearninggesturerecognition