Cargando…
A hierarchical sparse coding model predicts acoustic feature encoding in both auditory midbrain and cortex
The auditory pathway consists of multiple stages, from the cochlear nucleus to the auditory cortex. Neurons acting at different stages have different functions and exhibit different response properties. It is unclear whether these stages share a common encoding mechanism. We trained an unsupervised...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Public Library of Science
2019
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6386396/ https://www.ncbi.nlm.nih.gov/pubmed/30742609 http://dx.doi.org/10.1371/journal.pcbi.1006766 |
_version_ | 1783397376923271168 |
---|---|
author | Zhang, Qingtian Hu, Xiaolin Hong, Bo Zhang, Bo |
author_facet | Zhang, Qingtian Hu, Xiaolin Hong, Bo Zhang, Bo |
author_sort | Zhang, Qingtian |
collection | PubMed |
description | The auditory pathway consists of multiple stages, from the cochlear nucleus to the auditory cortex. Neurons acting at different stages have different functions and exhibit different response properties. It is unclear whether these stages share a common encoding mechanism. We trained an unsupervised deep learning model consisting of alternating sparse coding and max pooling layers on cochleogram-filtered human speech. Evaluation of the response properties revealed that computing units in lower layers exhibited spectro-temporal receptive fields (STRFs) similar to those of inferior colliculus neurons measured in physiological experiments, including properties such as sound onset and termination, checkerboard pattern, and spectral motion. Units in upper layers tended to be tuned to phonetic features such as plosivity and nasality, resembling the results of field recording in human auditory cortex. Variation of the sparseness level of the units in each higher layer revealed a positive correlation between the sparseness level and the strength of phonetic feature encoding. The activities of the units in the top layer, but not other layers, correlated with the dynamics of the first two formants (F1, F2) of all phonemes, indicating the encoding of phoneme dynamics in these units. These results suggest that the principles of sparse coding and max pooling may be universal in the human auditory pathway. |
format | Online Article Text |
id | pubmed-6386396 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2019 |
publisher | Public Library of Science |
record_format | MEDLINE/PubMed |
spelling | pubmed-63863962019-03-08 A hierarchical sparse coding model predicts acoustic feature encoding in both auditory midbrain and cortex Zhang, Qingtian Hu, Xiaolin Hong, Bo Zhang, Bo PLoS Comput Biol Research Article The auditory pathway consists of multiple stages, from the cochlear nucleus to the auditory cortex. Neurons acting at different stages have different functions and exhibit different response properties. It is unclear whether these stages share a common encoding mechanism. We trained an unsupervised deep learning model consisting of alternating sparse coding and max pooling layers on cochleogram-filtered human speech. Evaluation of the response properties revealed that computing units in lower layers exhibited spectro-temporal receptive fields (STRFs) similar to those of inferior colliculus neurons measured in physiological experiments, including properties such as sound onset and termination, checkerboard pattern, and spectral motion. Units in upper layers tended to be tuned to phonetic features such as plosivity and nasality, resembling the results of field recording in human auditory cortex. Variation of the sparseness level of the units in each higher layer revealed a positive correlation between the sparseness level and the strength of phonetic feature encoding. The activities of the units in the top layer, but not other layers, correlated with the dynamics of the first two formants (F1, F2) of all phonemes, indicating the encoding of phoneme dynamics in these units. These results suggest that the principles of sparse coding and max pooling may be universal in the human auditory pathway. Public Library of Science 2019-02-11 /pmc/articles/PMC6386396/ /pubmed/30742609 http://dx.doi.org/10.1371/journal.pcbi.1006766 Text en © 2019 Zhang et al http://creativecommons.org/licenses/by/4.0/ This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. |
spellingShingle | Research Article Zhang, Qingtian Hu, Xiaolin Hong, Bo Zhang, Bo A hierarchical sparse coding model predicts acoustic feature encoding in both auditory midbrain and cortex |
title | A hierarchical sparse coding model predicts acoustic feature encoding in both auditory midbrain and cortex |
title_full | A hierarchical sparse coding model predicts acoustic feature encoding in both auditory midbrain and cortex |
title_fullStr | A hierarchical sparse coding model predicts acoustic feature encoding in both auditory midbrain and cortex |
title_full_unstemmed | A hierarchical sparse coding model predicts acoustic feature encoding in both auditory midbrain and cortex |
title_short | A hierarchical sparse coding model predicts acoustic feature encoding in both auditory midbrain and cortex |
title_sort | hierarchical sparse coding model predicts acoustic feature encoding in both auditory midbrain and cortex |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6386396/ https://www.ncbi.nlm.nih.gov/pubmed/30742609 http://dx.doi.org/10.1371/journal.pcbi.1006766 |
work_keys_str_mv | AT zhangqingtian ahierarchicalsparsecodingmodelpredictsacousticfeatureencodinginbothauditorymidbrainandcortex AT huxiaolin ahierarchicalsparsecodingmodelpredictsacousticfeatureencodinginbothauditorymidbrainandcortex AT hongbo ahierarchicalsparsecodingmodelpredictsacousticfeatureencodinginbothauditorymidbrainandcortex AT zhangbo ahierarchicalsparsecodingmodelpredictsacousticfeatureencodinginbothauditorymidbrainandcortex AT zhangqingtian hierarchicalsparsecodingmodelpredictsacousticfeatureencodinginbothauditorymidbrainandcortex AT huxiaolin hierarchicalsparsecodingmodelpredictsacousticfeatureencodinginbothauditorymidbrainandcortex AT hongbo hierarchicalsparsecodingmodelpredictsacousticfeatureencodinginbothauditorymidbrainandcortex AT zhangbo hierarchicalsparsecodingmodelpredictsacousticfeatureencodinginbothauditorymidbrainandcortex |