Cargando…

A Gestalt inference model for auditory scene segregation

Our current understanding of how the brain segregates auditory scenes into meaningful objects is in line with a Gestaltism framework. These Gestalt principles suggest a theory of how different attributes of the soundscape are extracted then bound together into separate groups that reflect different...

Descripción completa

Detalles Bibliográficos
Autores principales: Chakrabarty, Debmalya, Elhilali, Mounya
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Public Library of Science 2019
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6358108/
https://www.ncbi.nlm.nih.gov/pubmed/30668568
http://dx.doi.org/10.1371/journal.pcbi.1006711
_version_ 1783391952005234688
author Chakrabarty, Debmalya
Elhilali, Mounya
author_facet Chakrabarty, Debmalya
Elhilali, Mounya
author_sort Chakrabarty, Debmalya
collection PubMed
description Our current understanding of how the brain segregates auditory scenes into meaningful objects is in line with a Gestaltism framework. These Gestalt principles suggest a theory of how different attributes of the soundscape are extracted then bound together into separate groups that reflect different objects or streams present in the scene. These cues are thought to reflect the underlying statistical structure of natural sounds in a similar way that statistics of natural images are closely linked to the principles that guide figure-ground segregation and object segmentation in vision. In the present study, we leverage inference in stochastic neural networks to learn emergent grouping cues directly from natural soundscapes including speech, music and sounds in nature. The model learns a hierarchy of local and global spectro-temporal attributes reminiscent of simultaneous and sequential Gestalt cues that underlie the organization of auditory scenes. These mappings operate at multiple time scales to analyze an incoming complex scene and are then fused using a Hebbian network that binds together coherent features into perceptually-segregated auditory objects. The proposed architecture successfully emulates a wide range of well established auditory scene segregation phenomena and quantifies the complimentary role of segregation and binding cues in driving auditory scene segregation.
format Online
Article
Text
id pubmed-6358108
institution National Center for Biotechnology Information
language English
publishDate 2019
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-63581082019-02-15 A Gestalt inference model for auditory scene segregation Chakrabarty, Debmalya Elhilali, Mounya PLoS Comput Biol Research Article Our current understanding of how the brain segregates auditory scenes into meaningful objects is in line with a Gestaltism framework. These Gestalt principles suggest a theory of how different attributes of the soundscape are extracted then bound together into separate groups that reflect different objects or streams present in the scene. These cues are thought to reflect the underlying statistical structure of natural sounds in a similar way that statistics of natural images are closely linked to the principles that guide figure-ground segregation and object segmentation in vision. In the present study, we leverage inference in stochastic neural networks to learn emergent grouping cues directly from natural soundscapes including speech, music and sounds in nature. The model learns a hierarchy of local and global spectro-temporal attributes reminiscent of simultaneous and sequential Gestalt cues that underlie the organization of auditory scenes. These mappings operate at multiple time scales to analyze an incoming complex scene and are then fused using a Hebbian network that binds together coherent features into perceptually-segregated auditory objects. The proposed architecture successfully emulates a wide range of well established auditory scene segregation phenomena and quantifies the complimentary role of segregation and binding cues in driving auditory scene segregation. Public Library of Science 2019-01-22 /pmc/articles/PMC6358108/ /pubmed/30668568 http://dx.doi.org/10.1371/journal.pcbi.1006711 Text en © 2019 Chakrabarty, Elhilali http://creativecommons.org/licenses/by/4.0/ This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
spellingShingle Research Article
Chakrabarty, Debmalya
Elhilali, Mounya
A Gestalt inference model for auditory scene segregation
title A Gestalt inference model for auditory scene segregation
title_full A Gestalt inference model for auditory scene segregation
title_fullStr A Gestalt inference model for auditory scene segregation
title_full_unstemmed A Gestalt inference model for auditory scene segregation
title_short A Gestalt inference model for auditory scene segregation
title_sort gestalt inference model for auditory scene segregation
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6358108/
https://www.ncbi.nlm.nih.gov/pubmed/30668568
http://dx.doi.org/10.1371/journal.pcbi.1006711
work_keys_str_mv AT chakrabartydebmalya agestaltinferencemodelforauditoryscenesegregation
AT elhilalimounya agestaltinferencemodelforauditoryscenesegregation
AT chakrabartydebmalya gestaltinferencemodelforauditoryscenesegregation
AT elhilalimounya gestaltinferencemodelforauditoryscenesegregation