Cargando…
A Bayesian computational basis for auditory selective attention using head rotation and the interaural time-difference cue
The process of resolving mixtures of several sounds into their separate individual streams is known as auditory scene analysis and it remains a challenging task for computational systems. It is well-known that animals use binaural differences in arrival time and intensity at the two ears to find the...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Public Library of Science
2017
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5629026/ https://www.ncbi.nlm.nih.gov/pubmed/28982139 http://dx.doi.org/10.1371/journal.pone.0186104 |
_version_ | 1783268986723500032 |
---|---|
author | Hambrook, Dillon A. Ilievski, Marko Mosadeghzad, Mohamad Tata, Matthew |
author_facet | Hambrook, Dillon A. Ilievski, Marko Mosadeghzad, Mohamad Tata, Matthew |
author_sort | Hambrook, Dillon A. |
collection | PubMed |
description | The process of resolving mixtures of several sounds into their separate individual streams is known as auditory scene analysis and it remains a challenging task for computational systems. It is well-known that animals use binaural differences in arrival time and intensity at the two ears to find the arrival angle of sounds in the azimuthal plane, and this localization function has sometimes been considered sufficient to enable the un-mixing of complex scenes. However, the ability of such systems to resolve distinct sound sources in both space and frequency remains limited. The neural computations for detecting interaural time difference (ITD) have been well studied and have served as the inspiration for computational auditory scene analysis systems, however a crucial limitation of ITD models is that they produce ambiguous or “phantom” images in the scene. This has been thought to limit their usefulness at frequencies above about 1khz in humans. We present a simple Bayesian model and an implementation on a robot that uses ITD information recursively. The model makes use of head rotations to show that ITD information is sufficient to unambiguously resolve sound sources in both space and frequency. Contrary to commonly held assumptions about sound localization, we show that the ITD cue used with high-frequency sound can provide accurate and unambiguous localization and resolution of competing sounds. Our findings suggest that an “active hearing” approach could be useful in robotic systems that operate in natural, noisy settings. We also suggest that neurophysiological models of sound localization in animals could benefit from revision to include the influence of top-down memory and sensorimotor integration across head rotations. |
format | Online Article Text |
id | pubmed-5629026 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2017 |
publisher | Public Library of Science |
record_format | MEDLINE/PubMed |
spelling | pubmed-56290262017-10-20 A Bayesian computational basis for auditory selective attention using head rotation and the interaural time-difference cue Hambrook, Dillon A. Ilievski, Marko Mosadeghzad, Mohamad Tata, Matthew PLoS One Research Article The process of resolving mixtures of several sounds into their separate individual streams is known as auditory scene analysis and it remains a challenging task for computational systems. It is well-known that animals use binaural differences in arrival time and intensity at the two ears to find the arrival angle of sounds in the azimuthal plane, and this localization function has sometimes been considered sufficient to enable the un-mixing of complex scenes. However, the ability of such systems to resolve distinct sound sources in both space and frequency remains limited. The neural computations for detecting interaural time difference (ITD) have been well studied and have served as the inspiration for computational auditory scene analysis systems, however a crucial limitation of ITD models is that they produce ambiguous or “phantom” images in the scene. This has been thought to limit their usefulness at frequencies above about 1khz in humans. We present a simple Bayesian model and an implementation on a robot that uses ITD information recursively. The model makes use of head rotations to show that ITD information is sufficient to unambiguously resolve sound sources in both space and frequency. Contrary to commonly held assumptions about sound localization, we show that the ITD cue used with high-frequency sound can provide accurate and unambiguous localization and resolution of competing sounds. Our findings suggest that an “active hearing” approach could be useful in robotic systems that operate in natural, noisy settings. We also suggest that neurophysiological models of sound localization in animals could benefit from revision to include the influence of top-down memory and sensorimotor integration across head rotations. Public Library of Science 2017-10-05 /pmc/articles/PMC5629026/ /pubmed/28982139 http://dx.doi.org/10.1371/journal.pone.0186104 Text en © 2017 Hambrook et al http://creativecommons.org/licenses/by/4.0/ This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. |
spellingShingle | Research Article Hambrook, Dillon A. Ilievski, Marko Mosadeghzad, Mohamad Tata, Matthew A Bayesian computational basis for auditory selective attention using head rotation and the interaural time-difference cue |
title | A Bayesian computational basis for auditory selective attention using head rotation and the interaural time-difference cue |
title_full | A Bayesian computational basis for auditory selective attention using head rotation and the interaural time-difference cue |
title_fullStr | A Bayesian computational basis for auditory selective attention using head rotation and the interaural time-difference cue |
title_full_unstemmed | A Bayesian computational basis for auditory selective attention using head rotation and the interaural time-difference cue |
title_short | A Bayesian computational basis for auditory selective attention using head rotation and the interaural time-difference cue |
title_sort | bayesian computational basis for auditory selective attention using head rotation and the interaural time-difference cue |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5629026/ https://www.ncbi.nlm.nih.gov/pubmed/28982139 http://dx.doi.org/10.1371/journal.pone.0186104 |
work_keys_str_mv | AT hambrookdillona abayesiancomputationalbasisforauditoryselectiveattentionusingheadrotationandtheinterauraltimedifferencecue AT ilievskimarko abayesiancomputationalbasisforauditoryselectiveattentionusingheadrotationandtheinterauraltimedifferencecue AT mosadeghzadmohamad abayesiancomputationalbasisforauditoryselectiveattentionusingheadrotationandtheinterauraltimedifferencecue AT tatamatthew abayesiancomputationalbasisforauditoryselectiveattentionusingheadrotationandtheinterauraltimedifferencecue AT hambrookdillona bayesiancomputationalbasisforauditoryselectiveattentionusingheadrotationandtheinterauraltimedifferencecue AT ilievskimarko bayesiancomputationalbasisforauditoryselectiveattentionusingheadrotationandtheinterauraltimedifferencecue AT mosadeghzadmohamad bayesiancomputationalbasisforauditoryselectiveattentionusingheadrotationandtheinterauraltimedifferencecue AT tatamatthew bayesiancomputationalbasisforauditoryselectiveattentionusingheadrotationandtheinterauraltimedifferencecue |