Cargando…

Connecting Deep Neural Networks to Physical, Perceptual, and Electrophysiological Auditory Signals

Deep neural networks have been recently shown to capture intricate information transformation of signals from the sensory profiles to semantic representations that facilitate recognition or discrimination of complex stimuli. In this vein, convolutional neural networks (CNNs) have been used very succ...

Descripción completa

Detalles Bibliográficos
Autores principales: Huang, Nicholas, Slaney, Malcolm, Elhilali, Mounya
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2018
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6102345/
https://www.ncbi.nlm.nih.gov/pubmed/30154688
http://dx.doi.org/10.3389/fnins.2018.00532
_version_ 1783349141765619712
author Huang, Nicholas
Slaney, Malcolm
Elhilali, Mounya
author_facet Huang, Nicholas
Slaney, Malcolm
Elhilali, Mounya
author_sort Huang, Nicholas
collection PubMed
description Deep neural networks have been recently shown to capture intricate information transformation of signals from the sensory profiles to semantic representations that facilitate recognition or discrimination of complex stimuli. In this vein, convolutional neural networks (CNNs) have been used very successfully in image and audio classification. Designed to imitate the hierarchical structure of the nervous system, CNNs reflect activation with increasing degrees of complexity that transform the incoming signal onto object-level representations. In this work, we employ a CNN trained for large-scale audio object classification to gain insights about the contribution of various audio representations that guide sound perception. The analysis contrasts activation of different layers of a CNN with acoustic features extracted directly from the scenes, perceptual salience obtained from behavioral responses of human listeners, as well as neural oscillations recorded by electroencephalography (EEG) in response to the same natural scenes. All three measures are tightly linked quantities believed to guide percepts of salience and object formation when listening to complex scenes. The results paint a picture of the intricate interplay between low-level and object-level representations in guiding auditory salience that is very much dependent on context and sound category.
format Online
Article
Text
id pubmed-6102345
institution National Center for Biotechnology Information
language English
publishDate 2018
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-61023452018-08-28 Connecting Deep Neural Networks to Physical, Perceptual, and Electrophysiological Auditory Signals Huang, Nicholas Slaney, Malcolm Elhilali, Mounya Front Neurosci Neuroscience Deep neural networks have been recently shown to capture intricate information transformation of signals from the sensory profiles to semantic representations that facilitate recognition or discrimination of complex stimuli. In this vein, convolutional neural networks (CNNs) have been used very successfully in image and audio classification. Designed to imitate the hierarchical structure of the nervous system, CNNs reflect activation with increasing degrees of complexity that transform the incoming signal onto object-level representations. In this work, we employ a CNN trained for large-scale audio object classification to gain insights about the contribution of various audio representations that guide sound perception. The analysis contrasts activation of different layers of a CNN with acoustic features extracted directly from the scenes, perceptual salience obtained from behavioral responses of human listeners, as well as neural oscillations recorded by electroencephalography (EEG) in response to the same natural scenes. All three measures are tightly linked quantities believed to guide percepts of salience and object formation when listening to complex scenes. The results paint a picture of the intricate interplay between low-level and object-level representations in guiding auditory salience that is very much dependent on context and sound category. Frontiers Media S.A. 2018-08-14 /pmc/articles/PMC6102345/ /pubmed/30154688 http://dx.doi.org/10.3389/fnins.2018.00532 Text en Copyright © 2018 Huang, Slaney and Elhilali. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Huang, Nicholas
Slaney, Malcolm
Elhilali, Mounya
Connecting Deep Neural Networks to Physical, Perceptual, and Electrophysiological Auditory Signals
title Connecting Deep Neural Networks to Physical, Perceptual, and Electrophysiological Auditory Signals
title_full Connecting Deep Neural Networks to Physical, Perceptual, and Electrophysiological Auditory Signals
title_fullStr Connecting Deep Neural Networks to Physical, Perceptual, and Electrophysiological Auditory Signals
title_full_unstemmed Connecting Deep Neural Networks to Physical, Perceptual, and Electrophysiological Auditory Signals
title_short Connecting Deep Neural Networks to Physical, Perceptual, and Electrophysiological Auditory Signals
title_sort connecting deep neural networks to physical, perceptual, and electrophysiological auditory signals
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6102345/
https://www.ncbi.nlm.nih.gov/pubmed/30154688
http://dx.doi.org/10.3389/fnins.2018.00532
work_keys_str_mv AT huangnicholas connectingdeepneuralnetworkstophysicalperceptualandelectrophysiologicalauditorysignals
AT slaneymalcolm connectingdeepneuralnetworkstophysicalperceptualandelectrophysiologicalauditorysignals
AT elhilalimounya connectingdeepneuralnetworkstophysicalperceptualandelectrophysiologicalauditorysignals