Cargando…

Modeling the Development of Audiovisual Cue Integration in Speech Perception

Adult speech perception is generally enhanced when information is provided from multiple modalities. In contrast, infants do not appear to benefit from combining auditory and visual speech information early in development. This is true despite the fact that both modalities are important to speech co...

Descripción completa

Detalles Bibliográficos
Autores principales: Getz, Laura M., Nordeen, Elke R., Vrabic, Sarah C., Toscano, Joseph C.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2017
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5366831/
https://www.ncbi.nlm.nih.gov/pubmed/28335558
http://dx.doi.org/10.3390/brainsci7030032
_version_ 1782517664202096640
author Getz, Laura M.
Nordeen, Elke R.
Vrabic, Sarah C.
Toscano, Joseph C.
author_facet Getz, Laura M.
Nordeen, Elke R.
Vrabic, Sarah C.
Toscano, Joseph C.
author_sort Getz, Laura M.
collection PubMed
description Adult speech perception is generally enhanced when information is provided from multiple modalities. In contrast, infants do not appear to benefit from combining auditory and visual speech information early in development. This is true despite the fact that both modalities are important to speech comprehension even at early stages of language acquisition. How then do listeners learn how to process auditory and visual information as part of a unified signal? In the auditory domain, statistical learning processes provide an excellent mechanism for acquiring phonological categories. Is this also true for the more complex problem of acquiring audiovisual correspondences, which require the learner to integrate information from multiple modalities? In this paper, we present simulations using Gaussian mixture models (GMMs) that learn cue weights and combine cues on the basis of their distributional statistics. First, we simulate the developmental process of acquiring phonological categories from auditory and visual cues, asking whether simple statistical learning approaches are sufficient for learning multi-modal representations. Second, we use this time course information to explain audiovisual speech perception in adult perceivers, including cases where auditory and visual input are mismatched. Overall, we find that domain-general statistical learning techniques allow us to model the developmental trajectory of audiovisual cue integration in speech, and in turn, allow us to better understand the mechanisms that give rise to unified percepts based on multiple cues.
format Online
Article
Text
id pubmed-5366831
institution National Center for Biotechnology Information
language English
publishDate 2017
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-53668312017-03-31 Modeling the Development of Audiovisual Cue Integration in Speech Perception Getz, Laura M. Nordeen, Elke R. Vrabic, Sarah C. Toscano, Joseph C. Brain Sci Article Adult speech perception is generally enhanced when information is provided from multiple modalities. In contrast, infants do not appear to benefit from combining auditory and visual speech information early in development. This is true despite the fact that both modalities are important to speech comprehension even at early stages of language acquisition. How then do listeners learn how to process auditory and visual information as part of a unified signal? In the auditory domain, statistical learning processes provide an excellent mechanism for acquiring phonological categories. Is this also true for the more complex problem of acquiring audiovisual correspondences, which require the learner to integrate information from multiple modalities? In this paper, we present simulations using Gaussian mixture models (GMMs) that learn cue weights and combine cues on the basis of their distributional statistics. First, we simulate the developmental process of acquiring phonological categories from auditory and visual cues, asking whether simple statistical learning approaches are sufficient for learning multi-modal representations. Second, we use this time course information to explain audiovisual speech perception in adult perceivers, including cases where auditory and visual input are mismatched. Overall, we find that domain-general statistical learning techniques allow us to model the developmental trajectory of audiovisual cue integration in speech, and in turn, allow us to better understand the mechanisms that give rise to unified percepts based on multiple cues. MDPI 2017-03-21 /pmc/articles/PMC5366831/ /pubmed/28335558 http://dx.doi.org/10.3390/brainsci7030032 Text en © 2017 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Getz, Laura M.
Nordeen, Elke R.
Vrabic, Sarah C.
Toscano, Joseph C.
Modeling the Development of Audiovisual Cue Integration in Speech Perception
title Modeling the Development of Audiovisual Cue Integration in Speech Perception
title_full Modeling the Development of Audiovisual Cue Integration in Speech Perception
title_fullStr Modeling the Development of Audiovisual Cue Integration in Speech Perception
title_full_unstemmed Modeling the Development of Audiovisual Cue Integration in Speech Perception
title_short Modeling the Development of Audiovisual Cue Integration in Speech Perception
title_sort modeling the development of audiovisual cue integration in speech perception
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5366831/
https://www.ncbi.nlm.nih.gov/pubmed/28335558
http://dx.doi.org/10.3390/brainsci7030032
work_keys_str_mv AT getzlauram modelingthedevelopmentofaudiovisualcueintegrationinspeechperception
AT nordeenelker modelingthedevelopmentofaudiovisualcueintegrationinspeechperception
AT vrabicsarahc modelingthedevelopmentofaudiovisualcueintegrationinspeechperception
AT toscanojosephc modelingthedevelopmentofaudiovisualcueintegrationinspeechperception