Cargando…

Humans Can Visually Judge Grasp Quality and Refine Their Judgments Through Visual and Haptic Feedback

How humans visually select where to grasp objects is determined by the physical object properties (e.g., size, shape, weight), the degrees of freedom of the arm and hand, as well as the task to be performed. We recently demonstrated that human grasps are near-optimal with respect to a weighted combi...

Descripción completa

Detalles Bibliográficos
Autores principales: Maiello, Guido, Schepko, Marcel, Klein, Lina K., Paulun, Vivian C., Fleming, Roland W.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7835720/
https://www.ncbi.nlm.nih.gov/pubmed/33510608
http://dx.doi.org/10.3389/fnins.2020.591898
_version_ 1783642587467350016
author Maiello, Guido
Schepko, Marcel
Klein, Lina K.
Paulun, Vivian C.
Fleming, Roland W.
author_facet Maiello, Guido
Schepko, Marcel
Klein, Lina K.
Paulun, Vivian C.
Fleming, Roland W.
author_sort Maiello, Guido
collection PubMed
description How humans visually select where to grasp objects is determined by the physical object properties (e.g., size, shape, weight), the degrees of freedom of the arm and hand, as well as the task to be performed. We recently demonstrated that human grasps are near-optimal with respect to a weighted combination of different cost functions that make grasps uncomfortable, unstable, or impossible, e.g., due to unnatural grasp apertures or large torques. Here, we ask whether humans can consciously access these rules. We test if humans can explicitly judge grasp quality derived from rules regarding grasp size, orientation, torque, and visibility. More specifically, we test if grasp quality can be inferred (i) by using visual cues and motor imagery alone, (ii) from watching grasps executed by others, and (iii) through performing grasps, i.e., receiving visual, proprioceptive and haptic feedback. Stimuli were novel objects made of 10 cubes of brass and wood (side length 2.5 cm) in various configurations. On each object, one near-optimal and one sub-optimal grasp were selected based on one cost function (e.g., torque), while the other constraints (grasp size, orientation, and visibility) were kept approximately constant or counterbalanced. Participants were visually cued to the location of the selected grasps on each object and verbally reported which of the two grasps was best. Across three experiments, participants were required to either (i) passively view the static objects and imagine executing the two competing grasps, (ii) passively view videos of other participants grasping the objects, or (iii) actively grasp the objects themselves. Our results show that, for a majority of tested objects, participants could already judge grasp optimality from simply viewing the objects and imagining to grasp them, but were significantly better in the video and grasping session. These findings suggest that humans can determine grasp quality even without performing the grasp—perhaps through motor imagery—and can further refine their understanding of how to correctly grasp an object through sensorimotor feedback but also by passively viewing others grasp objects.
format Online
Article
Text
id pubmed-7835720
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-78357202021-01-27 Humans Can Visually Judge Grasp Quality and Refine Their Judgments Through Visual and Haptic Feedback Maiello, Guido Schepko, Marcel Klein, Lina K. Paulun, Vivian C. Fleming, Roland W. Front Neurosci Neuroscience How humans visually select where to grasp objects is determined by the physical object properties (e.g., size, shape, weight), the degrees of freedom of the arm and hand, as well as the task to be performed. We recently demonstrated that human grasps are near-optimal with respect to a weighted combination of different cost functions that make grasps uncomfortable, unstable, or impossible, e.g., due to unnatural grasp apertures or large torques. Here, we ask whether humans can consciously access these rules. We test if humans can explicitly judge grasp quality derived from rules regarding grasp size, orientation, torque, and visibility. More specifically, we test if grasp quality can be inferred (i) by using visual cues and motor imagery alone, (ii) from watching grasps executed by others, and (iii) through performing grasps, i.e., receiving visual, proprioceptive and haptic feedback. Stimuli were novel objects made of 10 cubes of brass and wood (side length 2.5 cm) in various configurations. On each object, one near-optimal and one sub-optimal grasp were selected based on one cost function (e.g., torque), while the other constraints (grasp size, orientation, and visibility) were kept approximately constant or counterbalanced. Participants were visually cued to the location of the selected grasps on each object and verbally reported which of the two grasps was best. Across three experiments, participants were required to either (i) passively view the static objects and imagine executing the two competing grasps, (ii) passively view videos of other participants grasping the objects, or (iii) actively grasp the objects themselves. Our results show that, for a majority of tested objects, participants could already judge grasp optimality from simply viewing the objects and imagining to grasp them, but were significantly better in the video and grasping session. These findings suggest that humans can determine grasp quality even without performing the grasp—perhaps through motor imagery—and can further refine their understanding of how to correctly grasp an object through sensorimotor feedback but also by passively viewing others grasp objects. Frontiers Media S.A. 2021-01-12 /pmc/articles/PMC7835720/ /pubmed/33510608 http://dx.doi.org/10.3389/fnins.2020.591898 Text en Copyright © 2021 Maiello, Schepko, Klein, Paulun and Fleming. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Maiello, Guido
Schepko, Marcel
Klein, Lina K.
Paulun, Vivian C.
Fleming, Roland W.
Humans Can Visually Judge Grasp Quality and Refine Their Judgments Through Visual and Haptic Feedback
title Humans Can Visually Judge Grasp Quality and Refine Their Judgments Through Visual and Haptic Feedback
title_full Humans Can Visually Judge Grasp Quality and Refine Their Judgments Through Visual and Haptic Feedback
title_fullStr Humans Can Visually Judge Grasp Quality and Refine Their Judgments Through Visual and Haptic Feedback
title_full_unstemmed Humans Can Visually Judge Grasp Quality and Refine Their Judgments Through Visual and Haptic Feedback
title_short Humans Can Visually Judge Grasp Quality and Refine Their Judgments Through Visual and Haptic Feedback
title_sort humans can visually judge grasp quality and refine their judgments through visual and haptic feedback
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7835720/
https://www.ncbi.nlm.nih.gov/pubmed/33510608
http://dx.doi.org/10.3389/fnins.2020.591898
work_keys_str_mv AT maielloguido humanscanvisuallyjudgegraspqualityandrefinetheirjudgmentsthroughvisualandhapticfeedback
AT schepkomarcel humanscanvisuallyjudgegraspqualityandrefinetheirjudgmentsthroughvisualandhapticfeedback
AT kleinlinak humanscanvisuallyjudgegraspqualityandrefinetheirjudgmentsthroughvisualandhapticfeedback
AT paulunvivianc humanscanvisuallyjudgegraspqualityandrefinetheirjudgmentsthroughvisualandhapticfeedback
AT flemingrolandw humanscanvisuallyjudgegraspqualityandrefinetheirjudgmentsthroughvisualandhapticfeedback