Cargando…
Visual delay affects force scaling and weight perception during object lifting in virtual reality
Lifting an object requires precise scaling of fingertip forces based on a prediction of object weight. At object contact, a series of tactile and visual events arise that need to be rapidly processed online to fine-tune the planned motor commands for lifting the object. The brain mechanisms underlyi...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
American Physiological Society
2019
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6485735/ https://www.ncbi.nlm.nih.gov/pubmed/30673365 http://dx.doi.org/10.1152/jn.00396.2018 |
_version_ | 1783414293541158912 |
---|---|
author | van Polanen, Vonne Tibold, Robert Nuruki, Atsuo Davare, Marco |
author_facet | van Polanen, Vonne Tibold, Robert Nuruki, Atsuo Davare, Marco |
author_sort | van Polanen, Vonne |
collection | PubMed |
description | Lifting an object requires precise scaling of fingertip forces based on a prediction of object weight. At object contact, a series of tactile and visual events arise that need to be rapidly processed online to fine-tune the planned motor commands for lifting the object. The brain mechanisms underlying multisensory integration serially at transient sensorimotor events, a general feature of actions requiring hand-object interactions, are not yet understood. In this study we tested the relative weighting between haptic and visual signals when they are integrated online into the motor command. We used a new virtual reality setup to desynchronize visual feedback from haptics, which allowed us to probe the relative contribution of haptics and vision in driving participants’ movements when they grasped virtual objects simulated by two force-feedback robots. We found that visual delay changed the profile of fingertip force generation and led participants to perceive objects as heavier than when lifts were performed without visual delay. We further modeled the effect of vision on motor output by manipulating the extent to which delayed visual events could bias the force profile, which allowed us to determine the specific weighting the brain assigns to haptics and vision. Our results show for the first time how visuo-haptic integration is processed at discrete sensorimotor events for controlling object-lifting dynamics and further highlight the organization of multisensory signals online for controlling action and perception. NEW & NOTEWORTHY Dexterous hand movements require rapid integration of information from different senses, in particular touch and vision, at different key time points as movement unfolds. The relative weighting between vision and haptics for object manipulation is unknown. We used object lifting in virtual reality to desynchronize visual and haptic feedback and find out their relative weightings. Our findings shed light on how rapid multisensory integration is processed over a series of discrete sensorimotor control points. |
format | Online Article Text |
id | pubmed-6485735 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2019 |
publisher | American Physiological Society |
record_format | MEDLINE/PubMed |
spelling | pubmed-64857352020-04-01 Visual delay affects force scaling and weight perception during object lifting in virtual reality van Polanen, Vonne Tibold, Robert Nuruki, Atsuo Davare, Marco J Neurophysiol Research Article Lifting an object requires precise scaling of fingertip forces based on a prediction of object weight. At object contact, a series of tactile and visual events arise that need to be rapidly processed online to fine-tune the planned motor commands for lifting the object. The brain mechanisms underlying multisensory integration serially at transient sensorimotor events, a general feature of actions requiring hand-object interactions, are not yet understood. In this study we tested the relative weighting between haptic and visual signals when they are integrated online into the motor command. We used a new virtual reality setup to desynchronize visual feedback from haptics, which allowed us to probe the relative contribution of haptics and vision in driving participants’ movements when they grasped virtual objects simulated by two force-feedback robots. We found that visual delay changed the profile of fingertip force generation and led participants to perceive objects as heavier than when lifts were performed without visual delay. We further modeled the effect of vision on motor output by manipulating the extent to which delayed visual events could bias the force profile, which allowed us to determine the specific weighting the brain assigns to haptics and vision. Our results show for the first time how visuo-haptic integration is processed at discrete sensorimotor events for controlling object-lifting dynamics and further highlight the organization of multisensory signals online for controlling action and perception. NEW & NOTEWORTHY Dexterous hand movements require rapid integration of information from different senses, in particular touch and vision, at different key time points as movement unfolds. The relative weighting between vision and haptics for object manipulation is unknown. We used object lifting in virtual reality to desynchronize visual and haptic feedback and find out their relative weightings. Our findings shed light on how rapid multisensory integration is processed over a series of discrete sensorimotor control points. American Physiological Society 2019-04-01 2019-01-23 /pmc/articles/PMC6485735/ /pubmed/30673365 http://dx.doi.org/10.1152/jn.00396.2018 Text en Copyright © 2019 the American Physiological Society http://creativecommons.org/licenses/by/4.0/deed.en_US Licensed under Creative Commons Attribution CC-BY 4.0 (http://creativecommons.org/licenses/by/4.0/deed.en_US) : © the American Physiological Society. |
spellingShingle | Research Article van Polanen, Vonne Tibold, Robert Nuruki, Atsuo Davare, Marco Visual delay affects force scaling and weight perception during object lifting in virtual reality |
title | Visual delay affects force scaling and weight perception during object lifting in virtual reality |
title_full | Visual delay affects force scaling and weight perception during object lifting in virtual reality |
title_fullStr | Visual delay affects force scaling and weight perception during object lifting in virtual reality |
title_full_unstemmed | Visual delay affects force scaling and weight perception during object lifting in virtual reality |
title_short | Visual delay affects force scaling and weight perception during object lifting in virtual reality |
title_sort | visual delay affects force scaling and weight perception during object lifting in virtual reality |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6485735/ https://www.ncbi.nlm.nih.gov/pubmed/30673365 http://dx.doi.org/10.1152/jn.00396.2018 |
work_keys_str_mv | AT vanpolanenvonne visualdelayaffectsforcescalingandweightperceptionduringobjectliftinginvirtualreality AT tiboldrobert visualdelayaffectsforcescalingandweightperceptionduringobjectliftinginvirtualreality AT nurukiatsuo visualdelayaffectsforcescalingandweightperceptionduringobjectliftinginvirtualreality AT davaremarco visualdelayaffectsforcescalingandweightperceptionduringobjectliftinginvirtualreality |