Cargando…

A Bayesian computational basis for auditory selective attention using head rotation and the interaural time-difference cue

The process of resolving mixtures of several sounds into their separate individual streams is known as auditory scene analysis and it remains a challenging task for computational systems. It is well-known that animals use binaural differences in arrival time and intensity at the two ears to find the...

Descripción completa

Detalles Bibliográficos
Autores principales: Hambrook, Dillon A., Ilievski, Marko, Mosadeghzad, Mohamad, Tata, Matthew
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Public Library of Science 2017
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5629026/
https://www.ncbi.nlm.nih.gov/pubmed/28982139
http://dx.doi.org/10.1371/journal.pone.0186104
_version_ 1783268986723500032
author Hambrook, Dillon A.
Ilievski, Marko
Mosadeghzad, Mohamad
Tata, Matthew
author_facet Hambrook, Dillon A.
Ilievski, Marko
Mosadeghzad, Mohamad
Tata, Matthew
author_sort Hambrook, Dillon A.
collection PubMed
description The process of resolving mixtures of several sounds into their separate individual streams is known as auditory scene analysis and it remains a challenging task for computational systems. It is well-known that animals use binaural differences in arrival time and intensity at the two ears to find the arrival angle of sounds in the azimuthal plane, and this localization function has sometimes been considered sufficient to enable the un-mixing of complex scenes. However, the ability of such systems to resolve distinct sound sources in both space and frequency remains limited. The neural computations for detecting interaural time difference (ITD) have been well studied and have served as the inspiration for computational auditory scene analysis systems, however a crucial limitation of ITD models is that they produce ambiguous or “phantom” images in the scene. This has been thought to limit their usefulness at frequencies above about 1khz in humans. We present a simple Bayesian model and an implementation on a robot that uses ITD information recursively. The model makes use of head rotations to show that ITD information is sufficient to unambiguously resolve sound sources in both space and frequency. Contrary to commonly held assumptions about sound localization, we show that the ITD cue used with high-frequency sound can provide accurate and unambiguous localization and resolution of competing sounds. Our findings suggest that an “active hearing” approach could be useful in robotic systems that operate in natural, noisy settings. We also suggest that neurophysiological models of sound localization in animals could benefit from revision to include the influence of top-down memory and sensorimotor integration across head rotations.
format Online
Article
Text
id pubmed-5629026
institution National Center for Biotechnology Information
language English
publishDate 2017
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-56290262017-10-20 A Bayesian computational basis for auditory selective attention using head rotation and the interaural time-difference cue Hambrook, Dillon A. Ilievski, Marko Mosadeghzad, Mohamad Tata, Matthew PLoS One Research Article The process of resolving mixtures of several sounds into their separate individual streams is known as auditory scene analysis and it remains a challenging task for computational systems. It is well-known that animals use binaural differences in arrival time and intensity at the two ears to find the arrival angle of sounds in the azimuthal plane, and this localization function has sometimes been considered sufficient to enable the un-mixing of complex scenes. However, the ability of such systems to resolve distinct sound sources in both space and frequency remains limited. The neural computations for detecting interaural time difference (ITD) have been well studied and have served as the inspiration for computational auditory scene analysis systems, however a crucial limitation of ITD models is that they produce ambiguous or “phantom” images in the scene. This has been thought to limit their usefulness at frequencies above about 1khz in humans. We present a simple Bayesian model and an implementation on a robot that uses ITD information recursively. The model makes use of head rotations to show that ITD information is sufficient to unambiguously resolve sound sources in both space and frequency. Contrary to commonly held assumptions about sound localization, we show that the ITD cue used with high-frequency sound can provide accurate and unambiguous localization and resolution of competing sounds. Our findings suggest that an “active hearing” approach could be useful in robotic systems that operate in natural, noisy settings. We also suggest that neurophysiological models of sound localization in animals could benefit from revision to include the influence of top-down memory and sensorimotor integration across head rotations. Public Library of Science 2017-10-05 /pmc/articles/PMC5629026/ /pubmed/28982139 http://dx.doi.org/10.1371/journal.pone.0186104 Text en © 2017 Hambrook et al http://creativecommons.org/licenses/by/4.0/ This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
spellingShingle Research Article
Hambrook, Dillon A.
Ilievski, Marko
Mosadeghzad, Mohamad
Tata, Matthew
A Bayesian computational basis for auditory selective attention using head rotation and the interaural time-difference cue
title A Bayesian computational basis for auditory selective attention using head rotation and the interaural time-difference cue
title_full A Bayesian computational basis for auditory selective attention using head rotation and the interaural time-difference cue
title_fullStr A Bayesian computational basis for auditory selective attention using head rotation and the interaural time-difference cue
title_full_unstemmed A Bayesian computational basis for auditory selective attention using head rotation and the interaural time-difference cue
title_short A Bayesian computational basis for auditory selective attention using head rotation and the interaural time-difference cue
title_sort bayesian computational basis for auditory selective attention using head rotation and the interaural time-difference cue
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5629026/
https://www.ncbi.nlm.nih.gov/pubmed/28982139
http://dx.doi.org/10.1371/journal.pone.0186104
work_keys_str_mv AT hambrookdillona abayesiancomputationalbasisforauditoryselectiveattentionusingheadrotationandtheinterauraltimedifferencecue
AT ilievskimarko abayesiancomputationalbasisforauditoryselectiveattentionusingheadrotationandtheinterauraltimedifferencecue
AT mosadeghzadmohamad abayesiancomputationalbasisforauditoryselectiveattentionusingheadrotationandtheinterauraltimedifferencecue
AT tatamatthew abayesiancomputationalbasisforauditoryselectiveattentionusingheadrotationandtheinterauraltimedifferencecue
AT hambrookdillona bayesiancomputationalbasisforauditoryselectiveattentionusingheadrotationandtheinterauraltimedifferencecue
AT ilievskimarko bayesiancomputationalbasisforauditoryselectiveattentionusingheadrotationandtheinterauraltimedifferencecue
AT mosadeghzadmohamad bayesiancomputationalbasisforauditoryselectiveattentionusingheadrotationandtheinterauraltimedifferencecue
AT tatamatthew bayesiancomputationalbasisforauditoryselectiveattentionusingheadrotationandtheinterauraltimedifferencecue