Cargando…

Brains and algorithms partially converge in natural language processing

Deep learning algorithms trained to predict masked words from large amount of text have recently been shown to generate activations similar to those of the human brain. However, what drives this similarity remains currently unknown. Here, we systematically compare a variety of deep language models t...

Descripción completa

Detalles Bibliográficos
Autores principales: Caucheteux, Charlotte, King, Jean-Rémi
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Nature Publishing Group UK 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8850612/
https://www.ncbi.nlm.nih.gov/pubmed/35173264
http://dx.doi.org/10.1038/s42003-022-03036-1
_version_ 1784652636327247872
author Caucheteux, Charlotte
King, Jean-Rémi
author_facet Caucheteux, Charlotte
King, Jean-Rémi
author_sort Caucheteux, Charlotte
collection PubMed
description Deep learning algorithms trained to predict masked words from large amount of text have recently been shown to generate activations similar to those of the human brain. However, what drives this similarity remains currently unknown. Here, we systematically compare a variety of deep language models to identify the computational principles that lead them to generate brain-like representations of sentences. Specifically, we analyze the brain responses to 400 isolated sentences in a large cohort of 102 subjects, each recorded for two hours with functional magnetic resonance imaging (fMRI) and magnetoencephalography (MEG). We then test where and when each of these algorithms maps onto the brain responses. Finally, we estimate how the architecture, training, and performance of these models independently account for the generation of brain-like representations. Our analyses reveal two main findings. First, the similarity between the algorithms and the brain primarily depends on their ability to predict words from context. Second, this similarity reveals the rise and maintenance of perceptual, lexical, and compositional representations within each cortical region. Overall, this study shows that modern language algorithms partially converge towards brain-like solutions, and thus delineates a promising path to unravel the foundations of natural language processing.
format Online
Article
Text
id pubmed-8850612
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Nature Publishing Group UK
record_format MEDLINE/PubMed
spelling pubmed-88506122022-03-04 Brains and algorithms partially converge in natural language processing Caucheteux, Charlotte King, Jean-Rémi Commun Biol Article Deep learning algorithms trained to predict masked words from large amount of text have recently been shown to generate activations similar to those of the human brain. However, what drives this similarity remains currently unknown. Here, we systematically compare a variety of deep language models to identify the computational principles that lead them to generate brain-like representations of sentences. Specifically, we analyze the brain responses to 400 isolated sentences in a large cohort of 102 subjects, each recorded for two hours with functional magnetic resonance imaging (fMRI) and magnetoencephalography (MEG). We then test where and when each of these algorithms maps onto the brain responses. Finally, we estimate how the architecture, training, and performance of these models independently account for the generation of brain-like representations. Our analyses reveal two main findings. First, the similarity between the algorithms and the brain primarily depends on their ability to predict words from context. Second, this similarity reveals the rise and maintenance of perceptual, lexical, and compositional representations within each cortical region. Overall, this study shows that modern language algorithms partially converge towards brain-like solutions, and thus delineates a promising path to unravel the foundations of natural language processing. Nature Publishing Group UK 2022-02-16 /pmc/articles/PMC8850612/ /pubmed/35173264 http://dx.doi.org/10.1038/s42003-022-03036-1 Text en © The Author(s) 2022, corrected publication 2023 https://creativecommons.org/licenses/by/4.0/Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) .
spellingShingle Article
Caucheteux, Charlotte
King, Jean-Rémi
Brains and algorithms partially converge in natural language processing
title Brains and algorithms partially converge in natural language processing
title_full Brains and algorithms partially converge in natural language processing
title_fullStr Brains and algorithms partially converge in natural language processing
title_full_unstemmed Brains and algorithms partially converge in natural language processing
title_short Brains and algorithms partially converge in natural language processing
title_sort brains and algorithms partially converge in natural language processing
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8850612/
https://www.ncbi.nlm.nih.gov/pubmed/35173264
http://dx.doi.org/10.1038/s42003-022-03036-1
work_keys_str_mv AT caucheteuxcharlotte brainsandalgorithmspartiallyconvergeinnaturallanguageprocessing
AT kingjeanremi brainsandalgorithmspartiallyconvergeinnaturallanguageprocessing