Cargando…
Encoding of speech in convolutional layers and the brain stem based on language experience
Comparing artificial neural networks with outputs of neuroimaging techniques has recently seen substantial advances in (computer) vision and text-based language models. Here, we propose a framework to compare biological and artificial neural computations of spoken language representations and propos...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Nature Publishing Group UK
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10119295/ https://www.ncbi.nlm.nih.gov/pubmed/37081119 http://dx.doi.org/10.1038/s41598-023-33384-9 |
_version_ | 1785028994812346368 |
---|---|
author | Beguš, Gašper Zhou, Alan Zhao, T. Christina |
author_facet | Beguš, Gašper Zhou, Alan Zhao, T. Christina |
author_sort | Beguš, Gašper |
collection | PubMed |
description | Comparing artificial neural networks with outputs of neuroimaging techniques has recently seen substantial advances in (computer) vision and text-based language models. Here, we propose a framework to compare biological and artificial neural computations of spoken language representations and propose several new challenges to this paradigm. The proposed technique is based on a similar principle that underlies electroencephalography (EEG): averaging of neural (artificial or biological) activity across neurons in the time domain, and allows to compare encoding of any acoustic property in the brain and in intermediate convolutional layers of an artificial neural network. Our approach allows a direct comparison of responses to a phonetic property in the brain and in deep neural networks that requires no linear transformations between the signals. We argue that the brain stem response (cABR) and the response in intermediate convolutional layers to the exact same stimulus are highly similar without applying any transformations, and we quantify this observation. The proposed technique not only reveals similarities, but also allows for analysis of the encoding of actual acoustic properties in the two signals: we compare peak latency (i) in cABR relative to the stimulus in the brain stem and in (ii) intermediate convolutional layers relative to the input/output in deep convolutional networks. We also examine and compare the effect of prior language exposure on the peak latency in cABR and in intermediate convolutional layers. Substantial similarities in peak latency encoding between the human brain and intermediate convolutional networks emerge based on results from eight trained networks (including a replication experiment). The proposed technique can be used to compare encoding between the human brain and intermediate convolutional layers for any acoustic property and for other neuroimaging techniques. |
format | Online Article Text |
id | pubmed-10119295 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Nature Publishing Group UK |
record_format | MEDLINE/PubMed |
spelling | pubmed-101192952023-04-22 Encoding of speech in convolutional layers and the brain stem based on language experience Beguš, Gašper Zhou, Alan Zhao, T. Christina Sci Rep Article Comparing artificial neural networks with outputs of neuroimaging techniques has recently seen substantial advances in (computer) vision and text-based language models. Here, we propose a framework to compare biological and artificial neural computations of spoken language representations and propose several new challenges to this paradigm. The proposed technique is based on a similar principle that underlies electroencephalography (EEG): averaging of neural (artificial or biological) activity across neurons in the time domain, and allows to compare encoding of any acoustic property in the brain and in intermediate convolutional layers of an artificial neural network. Our approach allows a direct comparison of responses to a phonetic property in the brain and in deep neural networks that requires no linear transformations between the signals. We argue that the brain stem response (cABR) and the response in intermediate convolutional layers to the exact same stimulus are highly similar without applying any transformations, and we quantify this observation. The proposed technique not only reveals similarities, but also allows for analysis of the encoding of actual acoustic properties in the two signals: we compare peak latency (i) in cABR relative to the stimulus in the brain stem and in (ii) intermediate convolutional layers relative to the input/output in deep convolutional networks. We also examine and compare the effect of prior language exposure on the peak latency in cABR and in intermediate convolutional layers. Substantial similarities in peak latency encoding between the human brain and intermediate convolutional networks emerge based on results from eight trained networks (including a replication experiment). The proposed technique can be used to compare encoding between the human brain and intermediate convolutional layers for any acoustic property and for other neuroimaging techniques. Nature Publishing Group UK 2023-04-20 /pmc/articles/PMC10119295/ /pubmed/37081119 http://dx.doi.org/10.1038/s41598-023-33384-9 Text en © The Author(s) 2023 https://creativecommons.org/licenses/by/4.0/Open AccessThis article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) . |
spellingShingle | Article Beguš, Gašper Zhou, Alan Zhao, T. Christina Encoding of speech in convolutional layers and the brain stem based on language experience |
title | Encoding of speech in convolutional layers and the brain stem based on language experience |
title_full | Encoding of speech in convolutional layers and the brain stem based on language experience |
title_fullStr | Encoding of speech in convolutional layers and the brain stem based on language experience |
title_full_unstemmed | Encoding of speech in convolutional layers and the brain stem based on language experience |
title_short | Encoding of speech in convolutional layers and the brain stem based on language experience |
title_sort | encoding of speech in convolutional layers and the brain stem based on language experience |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10119295/ https://www.ncbi.nlm.nih.gov/pubmed/37081119 http://dx.doi.org/10.1038/s41598-023-33384-9 |
work_keys_str_mv | AT begusgasper encodingofspeechinconvolutionallayersandthebrainstembasedonlanguageexperience AT zhoualan encodingofspeechinconvolutionallayersandthebrainstembasedonlanguageexperience AT zhaotchristina encodingofspeechinconvolutionallayersandthebrainstembasedonlanguageexperience |