Cargando…

Visual delay affects force scaling and weight perception during object lifting in virtual reality

Lifting an object requires precise scaling of fingertip forces based on a prediction of object weight. At object contact, a series of tactile and visual events arise that need to be rapidly processed online to fine-tune the planned motor commands for lifting the object. The brain mechanisms underlyi...

Descripción completa

Detalles Bibliográficos
Autores principales: van Polanen, Vonne, Tibold, Robert, Nuruki, Atsuo, Davare, Marco
Formato: Online Artículo Texto
Lenguaje:English
Publicado: American Physiological Society 2019
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6485735/
https://www.ncbi.nlm.nih.gov/pubmed/30673365
http://dx.doi.org/10.1152/jn.00396.2018
_version_ 1783414293541158912
author van Polanen, Vonne
Tibold, Robert
Nuruki, Atsuo
Davare, Marco
author_facet van Polanen, Vonne
Tibold, Robert
Nuruki, Atsuo
Davare, Marco
author_sort van Polanen, Vonne
collection PubMed
description Lifting an object requires precise scaling of fingertip forces based on a prediction of object weight. At object contact, a series of tactile and visual events arise that need to be rapidly processed online to fine-tune the planned motor commands for lifting the object. The brain mechanisms underlying multisensory integration serially at transient sensorimotor events, a general feature of actions requiring hand-object interactions, are not yet understood. In this study we tested the relative weighting between haptic and visual signals when they are integrated online into the motor command. We used a new virtual reality setup to desynchronize visual feedback from haptics, which allowed us to probe the relative contribution of haptics and vision in driving participants’ movements when they grasped virtual objects simulated by two force-feedback robots. We found that visual delay changed the profile of fingertip force generation and led participants to perceive objects as heavier than when lifts were performed without visual delay. We further modeled the effect of vision on motor output by manipulating the extent to which delayed visual events could bias the force profile, which allowed us to determine the specific weighting the brain assigns to haptics and vision. Our results show for the first time how visuo-haptic integration is processed at discrete sensorimotor events for controlling object-lifting dynamics and further highlight the organization of multisensory signals online for controlling action and perception. NEW & NOTEWORTHY Dexterous hand movements require rapid integration of information from different senses, in particular touch and vision, at different key time points as movement unfolds. The relative weighting between vision and haptics for object manipulation is unknown. We used object lifting in virtual reality to desynchronize visual and haptic feedback and find out their relative weightings. Our findings shed light on how rapid multisensory integration is processed over a series of discrete sensorimotor control points.
format Online
Article
Text
id pubmed-6485735
institution National Center for Biotechnology Information
language English
publishDate 2019
publisher American Physiological Society
record_format MEDLINE/PubMed
spelling pubmed-64857352020-04-01 Visual delay affects force scaling and weight perception during object lifting in virtual reality van Polanen, Vonne Tibold, Robert Nuruki, Atsuo Davare, Marco J Neurophysiol Research Article Lifting an object requires precise scaling of fingertip forces based on a prediction of object weight. At object contact, a series of tactile and visual events arise that need to be rapidly processed online to fine-tune the planned motor commands for lifting the object. The brain mechanisms underlying multisensory integration serially at transient sensorimotor events, a general feature of actions requiring hand-object interactions, are not yet understood. In this study we tested the relative weighting between haptic and visual signals when they are integrated online into the motor command. We used a new virtual reality setup to desynchronize visual feedback from haptics, which allowed us to probe the relative contribution of haptics and vision in driving participants’ movements when they grasped virtual objects simulated by two force-feedback robots. We found that visual delay changed the profile of fingertip force generation and led participants to perceive objects as heavier than when lifts were performed without visual delay. We further modeled the effect of vision on motor output by manipulating the extent to which delayed visual events could bias the force profile, which allowed us to determine the specific weighting the brain assigns to haptics and vision. Our results show for the first time how visuo-haptic integration is processed at discrete sensorimotor events for controlling object-lifting dynamics and further highlight the organization of multisensory signals online for controlling action and perception. NEW & NOTEWORTHY Dexterous hand movements require rapid integration of information from different senses, in particular touch and vision, at different key time points as movement unfolds. The relative weighting between vision and haptics for object manipulation is unknown. We used object lifting in virtual reality to desynchronize visual and haptic feedback and find out their relative weightings. Our findings shed light on how rapid multisensory integration is processed over a series of discrete sensorimotor control points. American Physiological Society 2019-04-01 2019-01-23 /pmc/articles/PMC6485735/ /pubmed/30673365 http://dx.doi.org/10.1152/jn.00396.2018 Text en Copyright © 2019 the American Physiological Society http://creativecommons.org/licenses/by/4.0/deed.en_US Licensed under Creative Commons Attribution CC-BY 4.0 (http://creativecommons.org/licenses/by/4.0/deed.en_US) : © the American Physiological Society.
spellingShingle Research Article
van Polanen, Vonne
Tibold, Robert
Nuruki, Atsuo
Davare, Marco
Visual delay affects force scaling and weight perception during object lifting in virtual reality
title Visual delay affects force scaling and weight perception during object lifting in virtual reality
title_full Visual delay affects force scaling and weight perception during object lifting in virtual reality
title_fullStr Visual delay affects force scaling and weight perception during object lifting in virtual reality
title_full_unstemmed Visual delay affects force scaling and weight perception during object lifting in virtual reality
title_short Visual delay affects force scaling and weight perception during object lifting in virtual reality
title_sort visual delay affects force scaling and weight perception during object lifting in virtual reality
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6485735/
https://www.ncbi.nlm.nih.gov/pubmed/30673365
http://dx.doi.org/10.1152/jn.00396.2018
work_keys_str_mv AT vanpolanenvonne visualdelayaffectsforcescalingandweightperceptionduringobjectliftinginvirtualreality
AT tiboldrobert visualdelayaffectsforcescalingandweightperceptionduringobjectliftinginvirtualreality
AT nurukiatsuo visualdelayaffectsforcescalingandweightperceptionduringobjectliftinginvirtualreality
AT davaremarco visualdelayaffectsforcescalingandweightperceptionduringobjectliftinginvirtualreality