Cargando…

Cross-Modal Sensory Integration of Visual-Tactile Motion Information: Instrument Design and Human Psychophysics

Information obtained from multiple sensory modalities, such as vision and touch, is integrated to yield a holistic percept. As a haptic approach usually involves cross-modal sensory experiences, it is necessary to develop an apparatus that can characterize how a biological system integrates visual-t...

Descripción completa

Detalles Bibliográficos
Autores principales: Pei, Yu-Cheng, Chang, Ting-Yu, Lee, Tsung-Chi, Saha, Sudipta, Lai, Hsin-Yi, Gomez-Ramirez, Manuel, Chou, Shih-Wei, Wong, Alice M. K.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Molecular Diversity Preservation International (MDPI) 2013
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3715219/
https://www.ncbi.nlm.nih.gov/pubmed/23727955
http://dx.doi.org/10.3390/s130607212
_version_ 1782277419297669120
author Pei, Yu-Cheng
Chang, Ting-Yu
Lee, Tsung-Chi
Saha, Sudipta
Lai, Hsin-Yi
Gomez-Ramirez, Manuel
Chou, Shih-Wei
Wong, Alice M. K.
author_facet Pei, Yu-Cheng
Chang, Ting-Yu
Lee, Tsung-Chi
Saha, Sudipta
Lai, Hsin-Yi
Gomez-Ramirez, Manuel
Chou, Shih-Wei
Wong, Alice M. K.
author_sort Pei, Yu-Cheng
collection PubMed
description Information obtained from multiple sensory modalities, such as vision and touch, is integrated to yield a holistic percept. As a haptic approach usually involves cross-modal sensory experiences, it is necessary to develop an apparatus that can characterize how a biological system integrates visual-tactile sensory information as well as how a robotic device infers object information emanating from both vision and touch. In the present study, we develop a novel visual-tactile cross-modal integration stimulator that consists of an LED panel to present visual stimuli and a tactile stimulator with three degrees of freedom that can present tactile motion stimuli with arbitrary motion direction, speed, and indentation depth in the skin. The apparatus can present cross-modal stimuli in which the spatial locations of visual and tactile stimulations are perfectly aligned. We presented visual-tactile stimuli in which the visual and tactile directions were either congruent or incongruent, and human observers reported the perceived visual direction of motion. Results showed that perceived direction of visual motion can be biased by the direction of tactile motion when visual signals are weakened. The results also showed that the visual-tactile motion integration follows the rule of temporal congruency of multi-modal inputs, a fundamental property known for cross-modal integration.
format Online
Article
Text
id pubmed-3715219
institution National Center for Biotechnology Information
language English
publishDate 2013
publisher Molecular Diversity Preservation International (MDPI)
record_format MEDLINE/PubMed
spelling pubmed-37152192013-07-24 Cross-Modal Sensory Integration of Visual-Tactile Motion Information: Instrument Design and Human Psychophysics Pei, Yu-Cheng Chang, Ting-Yu Lee, Tsung-Chi Saha, Sudipta Lai, Hsin-Yi Gomez-Ramirez, Manuel Chou, Shih-Wei Wong, Alice M. K. Sensors (Basel) Article Information obtained from multiple sensory modalities, such as vision and touch, is integrated to yield a holistic percept. As a haptic approach usually involves cross-modal sensory experiences, it is necessary to develop an apparatus that can characterize how a biological system integrates visual-tactile sensory information as well as how a robotic device infers object information emanating from both vision and touch. In the present study, we develop a novel visual-tactile cross-modal integration stimulator that consists of an LED panel to present visual stimuli and a tactile stimulator with three degrees of freedom that can present tactile motion stimuli with arbitrary motion direction, speed, and indentation depth in the skin. The apparatus can present cross-modal stimuli in which the spatial locations of visual and tactile stimulations are perfectly aligned. We presented visual-tactile stimuli in which the visual and tactile directions were either congruent or incongruent, and human observers reported the perceived visual direction of motion. Results showed that perceived direction of visual motion can be biased by the direction of tactile motion when visual signals are weakened. The results also showed that the visual-tactile motion integration follows the rule of temporal congruency of multi-modal inputs, a fundamental property known for cross-modal integration. Molecular Diversity Preservation International (MDPI) 2013-05-31 /pmc/articles/PMC3715219/ /pubmed/23727955 http://dx.doi.org/10.3390/s130607212 Text en © 2013 by the authors; licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution license (http://creativecommons.org/licenses/by/3.0/).
spellingShingle Article
Pei, Yu-Cheng
Chang, Ting-Yu
Lee, Tsung-Chi
Saha, Sudipta
Lai, Hsin-Yi
Gomez-Ramirez, Manuel
Chou, Shih-Wei
Wong, Alice M. K.
Cross-Modal Sensory Integration of Visual-Tactile Motion Information: Instrument Design and Human Psychophysics
title Cross-Modal Sensory Integration of Visual-Tactile Motion Information: Instrument Design and Human Psychophysics
title_full Cross-Modal Sensory Integration of Visual-Tactile Motion Information: Instrument Design and Human Psychophysics
title_fullStr Cross-Modal Sensory Integration of Visual-Tactile Motion Information: Instrument Design and Human Psychophysics
title_full_unstemmed Cross-Modal Sensory Integration of Visual-Tactile Motion Information: Instrument Design and Human Psychophysics
title_short Cross-Modal Sensory Integration of Visual-Tactile Motion Information: Instrument Design and Human Psychophysics
title_sort cross-modal sensory integration of visual-tactile motion information: instrument design and human psychophysics
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3715219/
https://www.ncbi.nlm.nih.gov/pubmed/23727955
http://dx.doi.org/10.3390/s130607212
work_keys_str_mv AT peiyucheng crossmodalsensoryintegrationofvisualtactilemotioninformationinstrumentdesignandhumanpsychophysics
AT changtingyu crossmodalsensoryintegrationofvisualtactilemotioninformationinstrumentdesignandhumanpsychophysics
AT leetsungchi crossmodalsensoryintegrationofvisualtactilemotioninformationinstrumentdesignandhumanpsychophysics
AT sahasudipta crossmodalsensoryintegrationofvisualtactilemotioninformationinstrumentdesignandhumanpsychophysics
AT laihsinyi crossmodalsensoryintegrationofvisualtactilemotioninformationinstrumentdesignandhumanpsychophysics
AT gomezramirezmanuel crossmodalsensoryintegrationofvisualtactilemotioninformationinstrumentdesignandhumanpsychophysics
AT choushihwei crossmodalsensoryintegrationofvisualtactilemotioninformationinstrumentdesignandhumanpsychophysics
AT wongalicemk crossmodalsensoryintegrationofvisualtactilemotioninformationinstrumentdesignandhumanpsychophysics