Cargando…
Segregation of complex acoustic scenes based on temporal coherence
In contrast to the complex acoustic environments we encounter everyday, most studies of auditory segregation have used relatively simple signals. Here, we synthesized a new stimulus to examine the detection of coherent patterns (‘figures’) from overlapping ‘background’ signals. In a series of experi...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
eLife Sciences Publications, Ltd
2013
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3721234/ https://www.ncbi.nlm.nih.gov/pubmed/23898398 http://dx.doi.org/10.7554/eLife.00699 |
_version_ | 1782278051776692224 |
---|---|
author | Teki, Sundeep Chait, Maria Kumar, Sukhbinder Shamma, Shihab Griffiths, Timothy D |
author_facet | Teki, Sundeep Chait, Maria Kumar, Sukhbinder Shamma, Shihab Griffiths, Timothy D |
author_sort | Teki, Sundeep |
collection | PubMed |
description | In contrast to the complex acoustic environments we encounter everyday, most studies of auditory segregation have used relatively simple signals. Here, we synthesized a new stimulus to examine the detection of coherent patterns (‘figures’) from overlapping ‘background’ signals. In a series of experiments, we demonstrate that human listeners are remarkably sensitive to the emergence of such figures and can tolerate a variety of spectral and temporal perturbations. This robust behavior is consistent with the existence of automatic auditory segregation mechanisms that are highly sensitive to correlations across frequency and time. The observed behavior cannot be explained purely on the basis of adaptation-based models used to explain the segregation of deterministic narrowband signals. We show that the present results are consistent with the predictions of a model of auditory perceptual organization based on temporal coherence. Our data thus support a role for temporal coherence as an organizational principle underlying auditory segregation. DOI: http://dx.doi.org/10.7554/eLife.00699.001 |
format | Online Article Text |
id | pubmed-3721234 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2013 |
publisher | eLife Sciences Publications, Ltd |
record_format | MEDLINE/PubMed |
spelling | pubmed-37212342013-07-29 Segregation of complex acoustic scenes based on temporal coherence Teki, Sundeep Chait, Maria Kumar, Sukhbinder Shamma, Shihab Griffiths, Timothy D eLife Neuroscience In contrast to the complex acoustic environments we encounter everyday, most studies of auditory segregation have used relatively simple signals. Here, we synthesized a new stimulus to examine the detection of coherent patterns (‘figures’) from overlapping ‘background’ signals. In a series of experiments, we demonstrate that human listeners are remarkably sensitive to the emergence of such figures and can tolerate a variety of spectral and temporal perturbations. This robust behavior is consistent with the existence of automatic auditory segregation mechanisms that are highly sensitive to correlations across frequency and time. The observed behavior cannot be explained purely on the basis of adaptation-based models used to explain the segregation of deterministic narrowband signals. We show that the present results are consistent with the predictions of a model of auditory perceptual organization based on temporal coherence. Our data thus support a role for temporal coherence as an organizational principle underlying auditory segregation. DOI: http://dx.doi.org/10.7554/eLife.00699.001 eLife Sciences Publications, Ltd 2013-07-23 /pmc/articles/PMC3721234/ /pubmed/23898398 http://dx.doi.org/10.7554/eLife.00699 Text en Copyright © 2013, Teki et al http://creativecommons.org/licenses/by/3.0/ This article is distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/3.0/) , which permits unrestricted use and redistribution provided that the original author and source are credited. |
spellingShingle | Neuroscience Teki, Sundeep Chait, Maria Kumar, Sukhbinder Shamma, Shihab Griffiths, Timothy D Segregation of complex acoustic scenes based on temporal coherence |
title | Segregation of complex acoustic scenes based on temporal coherence |
title_full | Segregation of complex acoustic scenes based on temporal coherence |
title_fullStr | Segregation of complex acoustic scenes based on temporal coherence |
title_full_unstemmed | Segregation of complex acoustic scenes based on temporal coherence |
title_short | Segregation of complex acoustic scenes based on temporal coherence |
title_sort | segregation of complex acoustic scenes based on temporal coherence |
topic | Neuroscience |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3721234/ https://www.ncbi.nlm.nih.gov/pubmed/23898398 http://dx.doi.org/10.7554/eLife.00699 |
work_keys_str_mv | AT tekisundeep segregationofcomplexacousticscenesbasedontemporalcoherence AT chaitmaria segregationofcomplexacousticscenesbasedontemporalcoherence AT kumarsukhbinder segregationofcomplexacousticscenesbasedontemporalcoherence AT shammashihab segregationofcomplexacousticscenesbasedontemporalcoherence AT griffithstimothyd segregationofcomplexacousticscenesbasedontemporalcoherence |