Cargando…

Auditory, Visual and Audiovisual Speech Processing Streams in Superior Temporal Sulcus

The human superior temporal sulcus (STS) is responsive to visual and auditory information, including sounds and facial cues during speech recognition. We investigated the functional organization of STS with respect to modality-specific and multimodal speech representations. Twenty younger adult part...

Descripción completa

Detalles Bibliográficos
Autores principales: Venezia, Jonathan H., Vaden, Kenneth I., Rong, Feng, Maddox, Dale, Saberi, Kourosh, Hickok, Gregory
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2017
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5383672/
https://www.ncbi.nlm.nih.gov/pubmed/28439236
http://dx.doi.org/10.3389/fnhum.2017.00174
_version_ 1782520319123128320
author Venezia, Jonathan H.
Vaden, Kenneth I.
Rong, Feng
Maddox, Dale
Saberi, Kourosh
Hickok, Gregory
author_facet Venezia, Jonathan H.
Vaden, Kenneth I.
Rong, Feng
Maddox, Dale
Saberi, Kourosh
Hickok, Gregory
author_sort Venezia, Jonathan H.
collection PubMed
description The human superior temporal sulcus (STS) is responsive to visual and auditory information, including sounds and facial cues during speech recognition. We investigated the functional organization of STS with respect to modality-specific and multimodal speech representations. Twenty younger adult participants were instructed to perform an oddball detection task and were presented with auditory, visual, and audiovisual speech stimuli, as well as auditory and visual nonspeech control stimuli in a block fMRI design. Consistent with a hypothesized anterior-posterior processing gradient in STS, auditory, visual and audiovisual stimuli produced the largest BOLD effects in anterior, posterior and middle STS (mSTS), respectively, based on whole-brain, linear mixed effects and principal component analyses. Notably, the mSTS exhibited preferential responses to multisensory stimulation, as well as speech compared to nonspeech. Within the mid-posterior and mSTS regions, response preferences changed gradually from visual, to multisensory, to auditory moving posterior to anterior. Post hoc analysis of visual regions in the posterior STS revealed that a single subregion bordering the mSTS was insensitive to differences in low-level motion kinematics yet distinguished between visual speech and nonspeech based on multi-voxel activation patterns. These results suggest that auditory and visual speech representations are elaborated gradually within anterior and posterior processing streams, respectively, and may be integrated within the mSTS, which is sensitive to more abstract speech information within and across presentation modalities. The spatial organization of STS is consistent with processing streams that are hypothesized to synthesize perceptual speech representations from sensory signals that provide convergent information from visual and auditory modalities.
format Online
Article
Text
id pubmed-5383672
institution National Center for Biotechnology Information
language English
publishDate 2017
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-53836722017-04-24 Auditory, Visual and Audiovisual Speech Processing Streams in Superior Temporal Sulcus Venezia, Jonathan H. Vaden, Kenneth I. Rong, Feng Maddox, Dale Saberi, Kourosh Hickok, Gregory Front Hum Neurosci Neuroscience The human superior temporal sulcus (STS) is responsive to visual and auditory information, including sounds and facial cues during speech recognition. We investigated the functional organization of STS with respect to modality-specific and multimodal speech representations. Twenty younger adult participants were instructed to perform an oddball detection task and were presented with auditory, visual, and audiovisual speech stimuli, as well as auditory and visual nonspeech control stimuli in a block fMRI design. Consistent with a hypothesized anterior-posterior processing gradient in STS, auditory, visual and audiovisual stimuli produced the largest BOLD effects in anterior, posterior and middle STS (mSTS), respectively, based on whole-brain, linear mixed effects and principal component analyses. Notably, the mSTS exhibited preferential responses to multisensory stimulation, as well as speech compared to nonspeech. Within the mid-posterior and mSTS regions, response preferences changed gradually from visual, to multisensory, to auditory moving posterior to anterior. Post hoc analysis of visual regions in the posterior STS revealed that a single subregion bordering the mSTS was insensitive to differences in low-level motion kinematics yet distinguished between visual speech and nonspeech based on multi-voxel activation patterns. These results suggest that auditory and visual speech representations are elaborated gradually within anterior and posterior processing streams, respectively, and may be integrated within the mSTS, which is sensitive to more abstract speech information within and across presentation modalities. The spatial organization of STS is consistent with processing streams that are hypothesized to synthesize perceptual speech representations from sensory signals that provide convergent information from visual and auditory modalities. Frontiers Media S.A. 2017-04-07 /pmc/articles/PMC5383672/ /pubmed/28439236 http://dx.doi.org/10.3389/fnhum.2017.00174 Text en Copyright © 2017 Venezia, Vaden Jr. http://creativecommons.org/licenses/by/4.0/ , Rong, Maddox, Saberi and Hickok. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Venezia, Jonathan H.
Vaden, Kenneth I.
Rong, Feng
Maddox, Dale
Saberi, Kourosh
Hickok, Gregory
Auditory, Visual and Audiovisual Speech Processing Streams in Superior Temporal Sulcus
title Auditory, Visual and Audiovisual Speech Processing Streams in Superior Temporal Sulcus
title_full Auditory, Visual and Audiovisual Speech Processing Streams in Superior Temporal Sulcus
title_fullStr Auditory, Visual and Audiovisual Speech Processing Streams in Superior Temporal Sulcus
title_full_unstemmed Auditory, Visual and Audiovisual Speech Processing Streams in Superior Temporal Sulcus
title_short Auditory, Visual and Audiovisual Speech Processing Streams in Superior Temporal Sulcus
title_sort auditory, visual and audiovisual speech processing streams in superior temporal sulcus
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5383672/
https://www.ncbi.nlm.nih.gov/pubmed/28439236
http://dx.doi.org/10.3389/fnhum.2017.00174
work_keys_str_mv AT veneziajonathanh auditoryvisualandaudiovisualspeechprocessingstreamsinsuperiortemporalsulcus
AT vadenkennethi auditoryvisualandaudiovisualspeechprocessingstreamsinsuperiortemporalsulcus
AT rongfeng auditoryvisualandaudiovisualspeechprocessingstreamsinsuperiortemporalsulcus
AT maddoxdale auditoryvisualandaudiovisualspeechprocessingstreamsinsuperiortemporalsulcus
AT saberikourosh auditoryvisualandaudiovisualspeechprocessingstreamsinsuperiortemporalsulcus
AT hickokgregory auditoryvisualandaudiovisualspeechprocessingstreamsinsuperiortemporalsulcus