Cargando…

DeepDynamicHand: A Deep Neural Architecture for Labeling Hand Manipulation Strategies in Video Sources Exploiting Temporal Information

Humans are capable of complex manipulation interactions with the environment, relying on the intrinsic adaptability and compliance of their hands. Recently, soft robotic manipulation has attempted to reproduce such an extraordinary behavior, through the design of deformable yet robust end-effectors....

Descripción completa

Detalles Bibliográficos
Autores principales: Arapi, Visar, Della Santina, Cosimo, Bacciu, Davide, Bianchi, Matteo, Bicchi, Antonio
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2018
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6304372/
https://www.ncbi.nlm.nih.gov/pubmed/30618707
http://dx.doi.org/10.3389/fnbot.2018.00086
_version_ 1783382345089286144
author Arapi, Visar
Della Santina, Cosimo
Bacciu, Davide
Bianchi, Matteo
Bicchi, Antonio
author_facet Arapi, Visar
Della Santina, Cosimo
Bacciu, Davide
Bianchi, Matteo
Bicchi, Antonio
author_sort Arapi, Visar
collection PubMed
description Humans are capable of complex manipulation interactions with the environment, relying on the intrinsic adaptability and compliance of their hands. Recently, soft robotic manipulation has attempted to reproduce such an extraordinary behavior, through the design of deformable yet robust end-effectors. To this goal, the investigation of human behavior has become crucial to correctly inform technological developments of robotic hands that can successfully exploit environmental constraint as humans actually do. Among the different tools robotics can leverage on to achieve this objective, deep learning has emerged as a promising approach for the study and then the implementation of neuro-scientific observations on the artificial side. However, current approaches tend to neglect the dynamic nature of hand pose recognition problems, limiting the effectiveness of these techniques in identifying sequences of manipulation primitives underpinning action generation, e.g., during purposeful interaction with the environment. In this work, we propose a vision-based supervised Hand Pose Recognition method which, for the first time, takes into account temporal information to identify meaningful sequences of actions in grasping and manipulation tasks. More specifically, we apply Deep Neural Networks to automatically learn features from hand posture images that consist of frames extracted from grasping and manipulation task videos with objects and external environmental constraints. For training purposes, videos are divided into intervals, each associated to a specific action by a human supervisor. The proposed algorithm combines a Convolutional Neural Network to detect the hand within each video frame and a Recurrent Neural Network to predict the hand action in the current frame, while taking into consideration the history of actions performed in the previous frames. Experimental validation has been performed on two datasets of dynamic hand-centric strategies, where subjects regularly interact with objects and environment. Proposed architecture achieved a very good classification accuracy on both datasets, reaching performance up to 94%, and outperforming state of the art techniques. The outcomes of this study can be successfully applied to robotics, e.g., for planning and control of soft anthropomorphic manipulators.
format Online
Article
Text
id pubmed-6304372
institution National Center for Biotechnology Information
language English
publishDate 2018
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-63043722019-01-07 DeepDynamicHand: A Deep Neural Architecture for Labeling Hand Manipulation Strategies in Video Sources Exploiting Temporal Information Arapi, Visar Della Santina, Cosimo Bacciu, Davide Bianchi, Matteo Bicchi, Antonio Front Neurorobot Robotics and AI Humans are capable of complex manipulation interactions with the environment, relying on the intrinsic adaptability and compliance of their hands. Recently, soft robotic manipulation has attempted to reproduce such an extraordinary behavior, through the design of deformable yet robust end-effectors. To this goal, the investigation of human behavior has become crucial to correctly inform technological developments of robotic hands that can successfully exploit environmental constraint as humans actually do. Among the different tools robotics can leverage on to achieve this objective, deep learning has emerged as a promising approach for the study and then the implementation of neuro-scientific observations on the artificial side. However, current approaches tend to neglect the dynamic nature of hand pose recognition problems, limiting the effectiveness of these techniques in identifying sequences of manipulation primitives underpinning action generation, e.g., during purposeful interaction with the environment. In this work, we propose a vision-based supervised Hand Pose Recognition method which, for the first time, takes into account temporal information to identify meaningful sequences of actions in grasping and manipulation tasks. More specifically, we apply Deep Neural Networks to automatically learn features from hand posture images that consist of frames extracted from grasping and manipulation task videos with objects and external environmental constraints. For training purposes, videos are divided into intervals, each associated to a specific action by a human supervisor. The proposed algorithm combines a Convolutional Neural Network to detect the hand within each video frame and a Recurrent Neural Network to predict the hand action in the current frame, while taking into consideration the history of actions performed in the previous frames. Experimental validation has been performed on two datasets of dynamic hand-centric strategies, where subjects regularly interact with objects and environment. Proposed architecture achieved a very good classification accuracy on both datasets, reaching performance up to 94%, and outperforming state of the art techniques. The outcomes of this study can be successfully applied to robotics, e.g., for planning and control of soft anthropomorphic manipulators. Frontiers Media S.A. 2018-12-17 /pmc/articles/PMC6304372/ /pubmed/30618707 http://dx.doi.org/10.3389/fnbot.2018.00086 Text en Copyright © 2018 Arapi, Della Santina, Bacciu, Bianchi and Bicchi. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Robotics and AI
Arapi, Visar
Della Santina, Cosimo
Bacciu, Davide
Bianchi, Matteo
Bicchi, Antonio
DeepDynamicHand: A Deep Neural Architecture for Labeling Hand Manipulation Strategies in Video Sources Exploiting Temporal Information
title DeepDynamicHand: A Deep Neural Architecture for Labeling Hand Manipulation Strategies in Video Sources Exploiting Temporal Information
title_full DeepDynamicHand: A Deep Neural Architecture for Labeling Hand Manipulation Strategies in Video Sources Exploiting Temporal Information
title_fullStr DeepDynamicHand: A Deep Neural Architecture for Labeling Hand Manipulation Strategies in Video Sources Exploiting Temporal Information
title_full_unstemmed DeepDynamicHand: A Deep Neural Architecture for Labeling Hand Manipulation Strategies in Video Sources Exploiting Temporal Information
title_short DeepDynamicHand: A Deep Neural Architecture for Labeling Hand Manipulation Strategies in Video Sources Exploiting Temporal Information
title_sort deepdynamichand: a deep neural architecture for labeling hand manipulation strategies in video sources exploiting temporal information
topic Robotics and AI
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6304372/
https://www.ncbi.nlm.nih.gov/pubmed/30618707
http://dx.doi.org/10.3389/fnbot.2018.00086
work_keys_str_mv AT arapivisar deepdynamichandadeepneuralarchitectureforlabelinghandmanipulationstrategiesinvideosourcesexploitingtemporalinformation
AT dellasantinacosimo deepdynamichandadeepneuralarchitectureforlabelinghandmanipulationstrategiesinvideosourcesexploitingtemporalinformation
AT bacciudavide deepdynamichandadeepneuralarchitectureforlabelinghandmanipulationstrategiesinvideosourcesexploitingtemporalinformation
AT bianchimatteo deepdynamichandadeepneuralarchitectureforlabelinghandmanipulationstrategiesinvideosourcesexploitingtemporalinformation
AT bicchiantonio deepdynamichandadeepneuralarchitectureforlabelinghandmanipulationstrategiesinvideosourcesexploitingtemporalinformation