Cargando…

A hierarchical sparse coding model predicts acoustic feature encoding in both auditory midbrain and cortex

The auditory pathway consists of multiple stages, from the cochlear nucleus to the auditory cortex. Neurons acting at different stages have different functions and exhibit different response properties. It is unclear whether these stages share a common encoding mechanism. We trained an unsupervised...

Descripción completa

Detalles Bibliográficos
Autores principales: Zhang, Qingtian, Hu, Xiaolin, Hong, Bo, Zhang, Bo
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Public Library of Science 2019
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6386396/
https://www.ncbi.nlm.nih.gov/pubmed/30742609
http://dx.doi.org/10.1371/journal.pcbi.1006766
_version_ 1783397376923271168
author Zhang, Qingtian
Hu, Xiaolin
Hong, Bo
Zhang, Bo
author_facet Zhang, Qingtian
Hu, Xiaolin
Hong, Bo
Zhang, Bo
author_sort Zhang, Qingtian
collection PubMed
description The auditory pathway consists of multiple stages, from the cochlear nucleus to the auditory cortex. Neurons acting at different stages have different functions and exhibit different response properties. It is unclear whether these stages share a common encoding mechanism. We trained an unsupervised deep learning model consisting of alternating sparse coding and max pooling layers on cochleogram-filtered human speech. Evaluation of the response properties revealed that computing units in lower layers exhibited spectro-temporal receptive fields (STRFs) similar to those of inferior colliculus neurons measured in physiological experiments, including properties such as sound onset and termination, checkerboard pattern, and spectral motion. Units in upper layers tended to be tuned to phonetic features such as plosivity and nasality, resembling the results of field recording in human auditory cortex. Variation of the sparseness level of the units in each higher layer revealed a positive correlation between the sparseness level and the strength of phonetic feature encoding. The activities of the units in the top layer, but not other layers, correlated with the dynamics of the first two formants (F1, F2) of all phonemes, indicating the encoding of phoneme dynamics in these units. These results suggest that the principles of sparse coding and max pooling may be universal in the human auditory pathway.
format Online
Article
Text
id pubmed-6386396
institution National Center for Biotechnology Information
language English
publishDate 2019
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-63863962019-03-08 A hierarchical sparse coding model predicts acoustic feature encoding in both auditory midbrain and cortex Zhang, Qingtian Hu, Xiaolin Hong, Bo Zhang, Bo PLoS Comput Biol Research Article The auditory pathway consists of multiple stages, from the cochlear nucleus to the auditory cortex. Neurons acting at different stages have different functions and exhibit different response properties. It is unclear whether these stages share a common encoding mechanism. We trained an unsupervised deep learning model consisting of alternating sparse coding and max pooling layers on cochleogram-filtered human speech. Evaluation of the response properties revealed that computing units in lower layers exhibited spectro-temporal receptive fields (STRFs) similar to those of inferior colliculus neurons measured in physiological experiments, including properties such as sound onset and termination, checkerboard pattern, and spectral motion. Units in upper layers tended to be tuned to phonetic features such as plosivity and nasality, resembling the results of field recording in human auditory cortex. Variation of the sparseness level of the units in each higher layer revealed a positive correlation between the sparseness level and the strength of phonetic feature encoding. The activities of the units in the top layer, but not other layers, correlated with the dynamics of the first two formants (F1, F2) of all phonemes, indicating the encoding of phoneme dynamics in these units. These results suggest that the principles of sparse coding and max pooling may be universal in the human auditory pathway. Public Library of Science 2019-02-11 /pmc/articles/PMC6386396/ /pubmed/30742609 http://dx.doi.org/10.1371/journal.pcbi.1006766 Text en © 2019 Zhang et al http://creativecommons.org/licenses/by/4.0/ This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
spellingShingle Research Article
Zhang, Qingtian
Hu, Xiaolin
Hong, Bo
Zhang, Bo
A hierarchical sparse coding model predicts acoustic feature encoding in both auditory midbrain and cortex
title A hierarchical sparse coding model predicts acoustic feature encoding in both auditory midbrain and cortex
title_full A hierarchical sparse coding model predicts acoustic feature encoding in both auditory midbrain and cortex
title_fullStr A hierarchical sparse coding model predicts acoustic feature encoding in both auditory midbrain and cortex
title_full_unstemmed A hierarchical sparse coding model predicts acoustic feature encoding in both auditory midbrain and cortex
title_short A hierarchical sparse coding model predicts acoustic feature encoding in both auditory midbrain and cortex
title_sort hierarchical sparse coding model predicts acoustic feature encoding in both auditory midbrain and cortex
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6386396/
https://www.ncbi.nlm.nih.gov/pubmed/30742609
http://dx.doi.org/10.1371/journal.pcbi.1006766
work_keys_str_mv AT zhangqingtian ahierarchicalsparsecodingmodelpredictsacousticfeatureencodinginbothauditorymidbrainandcortex
AT huxiaolin ahierarchicalsparsecodingmodelpredictsacousticfeatureencodinginbothauditorymidbrainandcortex
AT hongbo ahierarchicalsparsecodingmodelpredictsacousticfeatureencodinginbothauditorymidbrainandcortex
AT zhangbo ahierarchicalsparsecodingmodelpredictsacousticfeatureencodinginbothauditorymidbrainandcortex
AT zhangqingtian hierarchicalsparsecodingmodelpredictsacousticfeatureencodinginbothauditorymidbrainandcortex
AT huxiaolin hierarchicalsparsecodingmodelpredictsacousticfeatureencodinginbothauditorymidbrainandcortex
AT hongbo hierarchicalsparsecodingmodelpredictsacousticfeatureencodinginbothauditorymidbrainandcortex
AT zhangbo hierarchicalsparsecodingmodelpredictsacousticfeatureencodinginbothauditorymidbrainandcortex