Cargando…
A mathematical theory of relational generalization in transitive inference
Humans and animals routinely infer relations between different items or events and generalize these relations to novel combinations of items (“compositional generalization”). This allows them to respond appropriately to radically novel circumstances and is fundamental to advanced cognition. However,...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Cold Spring Harbor Laboratory
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10473627/ https://www.ncbi.nlm.nih.gov/pubmed/37662223 http://dx.doi.org/10.1101/2023.08.22.554287 |
_version_ | 1785100310742564864 |
---|---|
author | Lippl, Samuel Kay, Kenneth Jensen, Greg Ferrera, Vincent P. Abbott, L.F. |
author_facet | Lippl, Samuel Kay, Kenneth Jensen, Greg Ferrera, Vincent P. Abbott, L.F. |
author_sort | Lippl, Samuel |
collection | PubMed |
description | Humans and animals routinely infer relations between different items or events and generalize these relations to novel combinations of items (“compositional generalization”). This allows them to respond appropriately to radically novel circumstances and is fundamental to advanced cognition. However, how learning systems (including the brain) can implement the necessary inductive biases has been unclear. Here we investigated transitive inference (TI), a classic relational task paradigm in which subjects must learn a relation (A > B and B > C) and generalize it to new combinations of items (A > C). Through mathematical analysis, we found that a broad range of biologically relevant learning models (e.g. gradient flow or ridge regression) perform TI successfully and recapitulate signature behavioral patterns long observed in living subjects. First, we found that models with item-wise additive representations automatically encode transitive relations. Second, for more general representations, a single scalar “conjunctivity factor” determines model behavior on TI and, further, the principle of norm minimization (a standard statistical inductive bias) enables models with fixed, partly conjunctive representations to generalize transitively. Finally, neural networks in the “rich regime,” which enables representation learning and often leads to better generalization, deviate in task behavior from living subjects and can make generalization errors. Our findings show systematically how minimal statistical learning principles can explain the rich behaviors empirically observed in TI in living subjects, uncover the mechanistic basis of transitive generalization in standard learning models, and lay out a formally tractable approach to understanding the neural basis of relational generalization. |
format | Online Article Text |
id | pubmed-10473627 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Cold Spring Harbor Laboratory |
record_format | MEDLINE/PubMed |
spelling | pubmed-104736272023-09-02 A mathematical theory of relational generalization in transitive inference Lippl, Samuel Kay, Kenneth Jensen, Greg Ferrera, Vincent P. Abbott, L.F. bioRxiv Article Humans and animals routinely infer relations between different items or events and generalize these relations to novel combinations of items (“compositional generalization”). This allows them to respond appropriately to radically novel circumstances and is fundamental to advanced cognition. However, how learning systems (including the brain) can implement the necessary inductive biases has been unclear. Here we investigated transitive inference (TI), a classic relational task paradigm in which subjects must learn a relation (A > B and B > C) and generalize it to new combinations of items (A > C). Through mathematical analysis, we found that a broad range of biologically relevant learning models (e.g. gradient flow or ridge regression) perform TI successfully and recapitulate signature behavioral patterns long observed in living subjects. First, we found that models with item-wise additive representations automatically encode transitive relations. Second, for more general representations, a single scalar “conjunctivity factor” determines model behavior on TI and, further, the principle of norm minimization (a standard statistical inductive bias) enables models with fixed, partly conjunctive representations to generalize transitively. Finally, neural networks in the “rich regime,” which enables representation learning and often leads to better generalization, deviate in task behavior from living subjects and can make generalization errors. Our findings show systematically how minimal statistical learning principles can explain the rich behaviors empirically observed in TI in living subjects, uncover the mechanistic basis of transitive generalization in standard learning models, and lay out a formally tractable approach to understanding the neural basis of relational generalization. Cold Spring Harbor Laboratory 2023-08-23 /pmc/articles/PMC10473627/ /pubmed/37662223 http://dx.doi.org/10.1101/2023.08.22.554287 Text en https://creativecommons.org/licenses/by-nc/4.0/This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License (https://creativecommons.org/licenses/by-nc/4.0/) , which allows reusers to distribute, remix, adapt, and build upon the material in any medium or format for noncommercial purposes only, and only so long as attribution is given to the creator. |
spellingShingle | Article Lippl, Samuel Kay, Kenneth Jensen, Greg Ferrera, Vincent P. Abbott, L.F. A mathematical theory of relational generalization in transitive inference |
title | A mathematical theory of relational generalization in transitive inference |
title_full | A mathematical theory of relational generalization in transitive inference |
title_fullStr | A mathematical theory of relational generalization in transitive inference |
title_full_unstemmed | A mathematical theory of relational generalization in transitive inference |
title_short | A mathematical theory of relational generalization in transitive inference |
title_sort | mathematical theory of relational generalization in transitive inference |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10473627/ https://www.ncbi.nlm.nih.gov/pubmed/37662223 http://dx.doi.org/10.1101/2023.08.22.554287 |
work_keys_str_mv | AT lipplsamuel amathematicaltheoryofrelationalgeneralizationintransitiveinference AT kaykenneth amathematicaltheoryofrelationalgeneralizationintransitiveinference AT jensengreg amathematicaltheoryofrelationalgeneralizationintransitiveinference AT ferreravincentp amathematicaltheoryofrelationalgeneralizationintransitiveinference AT abbottlf amathematicaltheoryofrelationalgeneralizationintransitiveinference AT lipplsamuel mathematicaltheoryofrelationalgeneralizationintransitiveinference AT kaykenneth mathematicaltheoryofrelationalgeneralizationintransitiveinference AT jensengreg mathematicaltheoryofrelationalgeneralizationintransitiveinference AT ferreravincentp mathematicaltheoryofrelationalgeneralizationintransitiveinference AT abbottlf mathematicaltheoryofrelationalgeneralizationintransitiveinference |