Cargando…
Adaptive Local Spatiotemporal Features from RGB-D Data for One-Shot Learning Gesture Recognition
Noise and constant empirical motion constraints affect the extraction of distinctive spatiotemporal features from one or a few samples per gesture class. To tackle these problems, an adaptive local spatiotemporal feature (ALSTF) using fused RGB-D data is proposed. First, motion regions of interest (...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2016
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5191150/ https://www.ncbi.nlm.nih.gov/pubmed/27999337 http://dx.doi.org/10.3390/s16122171 |
_version_ | 1782487567775563776 |
---|---|
author | Lin, Jia Ruan, Xiaogang Yu, Naigong Yang, Yee-Hong |
author_facet | Lin, Jia Ruan, Xiaogang Yu, Naigong Yang, Yee-Hong |
author_sort | Lin, Jia |
collection | PubMed |
description | Noise and constant empirical motion constraints affect the extraction of distinctive spatiotemporal features from one or a few samples per gesture class. To tackle these problems, an adaptive local spatiotemporal feature (ALSTF) using fused RGB-D data is proposed. First, motion regions of interest (MRoIs) are adaptively extracted using grayscale and depth velocity variance information to greatly reduce the impact of noise. Then, corners are used as keypoints if their depth, and velocities of grayscale and of depth meet several adaptive local constraints in each MRoI. With further filtering of noise, an accurate and sufficient number of keypoints is obtained within the desired moving body parts (MBPs). Finally, four kinds of multiple descriptors are calculated and combined in extended gradient and motion spaces to represent the appearance and motion features of gestures. The experimental results on the ChaLearn gesture, CAD-60 and MSRDailyActivity3D datasets demonstrate that the proposed feature achieves higher performance compared with published state-of-the-art approaches under the one-shot learning setting and comparable accuracy under the leave-one-out cross validation. |
format | Online Article Text |
id | pubmed-5191150 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2016 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-51911502017-01-03 Adaptive Local Spatiotemporal Features from RGB-D Data for One-Shot Learning Gesture Recognition Lin, Jia Ruan, Xiaogang Yu, Naigong Yang, Yee-Hong Sensors (Basel) Article Noise and constant empirical motion constraints affect the extraction of distinctive spatiotemporal features from one or a few samples per gesture class. To tackle these problems, an adaptive local spatiotemporal feature (ALSTF) using fused RGB-D data is proposed. First, motion regions of interest (MRoIs) are adaptively extracted using grayscale and depth velocity variance information to greatly reduce the impact of noise. Then, corners are used as keypoints if their depth, and velocities of grayscale and of depth meet several adaptive local constraints in each MRoI. With further filtering of noise, an accurate and sufficient number of keypoints is obtained within the desired moving body parts (MBPs). Finally, four kinds of multiple descriptors are calculated and combined in extended gradient and motion spaces to represent the appearance and motion features of gestures. The experimental results on the ChaLearn gesture, CAD-60 and MSRDailyActivity3D datasets demonstrate that the proposed feature achieves higher performance compared with published state-of-the-art approaches under the one-shot learning setting and comparable accuracy under the leave-one-out cross validation. MDPI 2016-12-17 /pmc/articles/PMC5191150/ /pubmed/27999337 http://dx.doi.org/10.3390/s16122171 Text en © 2016 by the authors; licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC-BY) license (http://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Lin, Jia Ruan, Xiaogang Yu, Naigong Yang, Yee-Hong Adaptive Local Spatiotemporal Features from RGB-D Data for One-Shot Learning Gesture Recognition |
title | Adaptive Local Spatiotemporal Features from RGB-D Data for One-Shot Learning Gesture Recognition |
title_full | Adaptive Local Spatiotemporal Features from RGB-D Data for One-Shot Learning Gesture Recognition |
title_fullStr | Adaptive Local Spatiotemporal Features from RGB-D Data for One-Shot Learning Gesture Recognition |
title_full_unstemmed | Adaptive Local Spatiotemporal Features from RGB-D Data for One-Shot Learning Gesture Recognition |
title_short | Adaptive Local Spatiotemporal Features from RGB-D Data for One-Shot Learning Gesture Recognition |
title_sort | adaptive local spatiotemporal features from rgb-d data for one-shot learning gesture recognition |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5191150/ https://www.ncbi.nlm.nih.gov/pubmed/27999337 http://dx.doi.org/10.3390/s16122171 |
work_keys_str_mv | AT linjia adaptivelocalspatiotemporalfeaturesfromrgbddataforoneshotlearninggesturerecognition AT ruanxiaogang adaptivelocalspatiotemporalfeaturesfromrgbddataforoneshotlearninggesturerecognition AT yunaigong adaptivelocalspatiotemporalfeaturesfromrgbddataforoneshotlearninggesturerecognition AT yangyeehong adaptivelocalspatiotemporalfeaturesfromrgbddataforoneshotlearninggesturerecognition |