Cargando…
Connecting Deep Neural Networks to Physical, Perceptual, and Electrophysiological Auditory Signals
Deep neural networks have been recently shown to capture intricate information transformation of signals from the sensory profiles to semantic representations that facilitate recognition or discrimination of complex stimuli. In this vein, convolutional neural networks (CNNs) have been used very succ...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2018
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6102345/ https://www.ncbi.nlm.nih.gov/pubmed/30154688 http://dx.doi.org/10.3389/fnins.2018.00532 |
_version_ | 1783349141765619712 |
---|---|
author | Huang, Nicholas Slaney, Malcolm Elhilali, Mounya |
author_facet | Huang, Nicholas Slaney, Malcolm Elhilali, Mounya |
author_sort | Huang, Nicholas |
collection | PubMed |
description | Deep neural networks have been recently shown to capture intricate information transformation of signals from the sensory profiles to semantic representations that facilitate recognition or discrimination of complex stimuli. In this vein, convolutional neural networks (CNNs) have been used very successfully in image and audio classification. Designed to imitate the hierarchical structure of the nervous system, CNNs reflect activation with increasing degrees of complexity that transform the incoming signal onto object-level representations. In this work, we employ a CNN trained for large-scale audio object classification to gain insights about the contribution of various audio representations that guide sound perception. The analysis contrasts activation of different layers of a CNN with acoustic features extracted directly from the scenes, perceptual salience obtained from behavioral responses of human listeners, as well as neural oscillations recorded by electroencephalography (EEG) in response to the same natural scenes. All three measures are tightly linked quantities believed to guide percepts of salience and object formation when listening to complex scenes. The results paint a picture of the intricate interplay between low-level and object-level representations in guiding auditory salience that is very much dependent on context and sound category. |
format | Online Article Text |
id | pubmed-6102345 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2018 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-61023452018-08-28 Connecting Deep Neural Networks to Physical, Perceptual, and Electrophysiological Auditory Signals Huang, Nicholas Slaney, Malcolm Elhilali, Mounya Front Neurosci Neuroscience Deep neural networks have been recently shown to capture intricate information transformation of signals from the sensory profiles to semantic representations that facilitate recognition or discrimination of complex stimuli. In this vein, convolutional neural networks (CNNs) have been used very successfully in image and audio classification. Designed to imitate the hierarchical structure of the nervous system, CNNs reflect activation with increasing degrees of complexity that transform the incoming signal onto object-level representations. In this work, we employ a CNN trained for large-scale audio object classification to gain insights about the contribution of various audio representations that guide sound perception. The analysis contrasts activation of different layers of a CNN with acoustic features extracted directly from the scenes, perceptual salience obtained from behavioral responses of human listeners, as well as neural oscillations recorded by electroencephalography (EEG) in response to the same natural scenes. All three measures are tightly linked quantities believed to guide percepts of salience and object formation when listening to complex scenes. The results paint a picture of the intricate interplay between low-level and object-level representations in guiding auditory salience that is very much dependent on context and sound category. Frontiers Media S.A. 2018-08-14 /pmc/articles/PMC6102345/ /pubmed/30154688 http://dx.doi.org/10.3389/fnins.2018.00532 Text en Copyright © 2018 Huang, Slaney and Elhilali. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Neuroscience Huang, Nicholas Slaney, Malcolm Elhilali, Mounya Connecting Deep Neural Networks to Physical, Perceptual, and Electrophysiological Auditory Signals |
title | Connecting Deep Neural Networks to Physical, Perceptual, and Electrophysiological Auditory Signals |
title_full | Connecting Deep Neural Networks to Physical, Perceptual, and Electrophysiological Auditory Signals |
title_fullStr | Connecting Deep Neural Networks to Physical, Perceptual, and Electrophysiological Auditory Signals |
title_full_unstemmed | Connecting Deep Neural Networks to Physical, Perceptual, and Electrophysiological Auditory Signals |
title_short | Connecting Deep Neural Networks to Physical, Perceptual, and Electrophysiological Auditory Signals |
title_sort | connecting deep neural networks to physical, perceptual, and electrophysiological auditory signals |
topic | Neuroscience |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6102345/ https://www.ncbi.nlm.nih.gov/pubmed/30154688 http://dx.doi.org/10.3389/fnins.2018.00532 |
work_keys_str_mv | AT huangnicholas connectingdeepneuralnetworkstophysicalperceptualandelectrophysiologicalauditorysignals AT slaneymalcolm connectingdeepneuralnetworkstophysicalperceptualandelectrophysiologicalauditorysignals AT elhilalimounya connectingdeepneuralnetworkstophysicalperceptualandelectrophysiologicalauditorysignals |