Cargando…

Segregation of complex acoustic scenes based on temporal coherence

In contrast to the complex acoustic environments we encounter everyday, most studies of auditory segregation have used relatively simple signals. Here, we synthesized a new stimulus to examine the detection of coherent patterns (‘figures’) from overlapping ‘background’ signals. In a series of experi...

Descripción completa

Detalles Bibliográficos
Autores principales: Teki, Sundeep, Chait, Maria, Kumar, Sukhbinder, Shamma, Shihab, Griffiths, Timothy D
Formato: Online Artículo Texto
Lenguaje:English
Publicado: eLife Sciences Publications, Ltd 2013
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3721234/
https://www.ncbi.nlm.nih.gov/pubmed/23898398
http://dx.doi.org/10.7554/eLife.00699
_version_ 1782278051776692224
author Teki, Sundeep
Chait, Maria
Kumar, Sukhbinder
Shamma, Shihab
Griffiths, Timothy D
author_facet Teki, Sundeep
Chait, Maria
Kumar, Sukhbinder
Shamma, Shihab
Griffiths, Timothy D
author_sort Teki, Sundeep
collection PubMed
description In contrast to the complex acoustic environments we encounter everyday, most studies of auditory segregation have used relatively simple signals. Here, we synthesized a new stimulus to examine the detection of coherent patterns (‘figures’) from overlapping ‘background’ signals. In a series of experiments, we demonstrate that human listeners are remarkably sensitive to the emergence of such figures and can tolerate a variety of spectral and temporal perturbations. This robust behavior is consistent with the existence of automatic auditory segregation mechanisms that are highly sensitive to correlations across frequency and time. The observed behavior cannot be explained purely on the basis of adaptation-based models used to explain the segregation of deterministic narrowband signals. We show that the present results are consistent with the predictions of a model of auditory perceptual organization based on temporal coherence. Our data thus support a role for temporal coherence as an organizational principle underlying auditory segregation. DOI: http://dx.doi.org/10.7554/eLife.00699.001
format Online
Article
Text
id pubmed-3721234
institution National Center for Biotechnology Information
language English
publishDate 2013
publisher eLife Sciences Publications, Ltd
record_format MEDLINE/PubMed
spelling pubmed-37212342013-07-29 Segregation of complex acoustic scenes based on temporal coherence Teki, Sundeep Chait, Maria Kumar, Sukhbinder Shamma, Shihab Griffiths, Timothy D eLife Neuroscience In contrast to the complex acoustic environments we encounter everyday, most studies of auditory segregation have used relatively simple signals. Here, we synthesized a new stimulus to examine the detection of coherent patterns (‘figures’) from overlapping ‘background’ signals. In a series of experiments, we demonstrate that human listeners are remarkably sensitive to the emergence of such figures and can tolerate a variety of spectral and temporal perturbations. This robust behavior is consistent with the existence of automatic auditory segregation mechanisms that are highly sensitive to correlations across frequency and time. The observed behavior cannot be explained purely on the basis of adaptation-based models used to explain the segregation of deterministic narrowband signals. We show that the present results are consistent with the predictions of a model of auditory perceptual organization based on temporal coherence. Our data thus support a role for temporal coherence as an organizational principle underlying auditory segregation. DOI: http://dx.doi.org/10.7554/eLife.00699.001 eLife Sciences Publications, Ltd 2013-07-23 /pmc/articles/PMC3721234/ /pubmed/23898398 http://dx.doi.org/10.7554/eLife.00699 Text en Copyright © 2013, Teki et al http://creativecommons.org/licenses/by/3.0/ This article is distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/3.0/) , which permits unrestricted use and redistribution provided that the original author and source are credited.
spellingShingle Neuroscience
Teki, Sundeep
Chait, Maria
Kumar, Sukhbinder
Shamma, Shihab
Griffiths, Timothy D
Segregation of complex acoustic scenes based on temporal coherence
title Segregation of complex acoustic scenes based on temporal coherence
title_full Segregation of complex acoustic scenes based on temporal coherence
title_fullStr Segregation of complex acoustic scenes based on temporal coherence
title_full_unstemmed Segregation of complex acoustic scenes based on temporal coherence
title_short Segregation of complex acoustic scenes based on temporal coherence
title_sort segregation of complex acoustic scenes based on temporal coherence
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3721234/
https://www.ncbi.nlm.nih.gov/pubmed/23898398
http://dx.doi.org/10.7554/eLife.00699
work_keys_str_mv AT tekisundeep segregationofcomplexacousticscenesbasedontemporalcoherence
AT chaitmaria segregationofcomplexacousticscenesbasedontemporalcoherence
AT kumarsukhbinder segregationofcomplexacousticscenesbasedontemporalcoherence
AT shammashihab segregationofcomplexacousticscenesbasedontemporalcoherence
AT griffithstimothyd segregationofcomplexacousticscenesbasedontemporalcoherence