Cargando…

Emergence of analogy from relation learning

By middle childhood, humans are able to learn abstract semantic relations (e.g., antonym, synonym, category membership) and use them to reason by analogy. A deep theoretical challenge is to show how such abstract relations can arise from nonrelational inputs, thereby providing key elements of a prot...

Descripción completa

Detalles Bibliográficos
Autores principales: Lu, Hongjing, Wu, Ying Nian, Holyoak, Keith J.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: National Academy of Sciences 2019
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6410800/
https://www.ncbi.nlm.nih.gov/pubmed/30770443
http://dx.doi.org/10.1073/pnas.1814779116
_version_ 1783402313076965376
author Lu, Hongjing
Wu, Ying Nian
Holyoak, Keith J.
author_facet Lu, Hongjing
Wu, Ying Nian
Holyoak, Keith J.
author_sort Lu, Hongjing
collection PubMed
description By middle childhood, humans are able to learn abstract semantic relations (e.g., antonym, synonym, category membership) and use them to reason by analogy. A deep theoretical challenge is to show how such abstract relations can arise from nonrelational inputs, thereby providing key elements of a protosymbolic representation system. We have developed a computational model that exploits the potential synergy between deep learning from “big data” (to create semantic features for individual words) and supervised learning from “small data” (to create representations of semantic relations between words). Given as inputs labeled pairs of lexical representations extracted by deep learning, the model creates augmented representations by remapping features according to the rank of differences between values for the two words in each pair. These augmented representations aid in coping with the feature alignment problem (e.g., matching those features that make “love-hate” an antonym with the different features that make “rich-poor” an antonym). The model extracts weight distributions that are used to estimate the probabilities that new word pairs instantiate each relation, capturing the pattern of human typicality judgments for a broad range of abstract semantic relations. A measure of relational similarity can be derived and used to solve simple verbal analogies with human-level accuracy. Because each acquired relation has a modular representation, basic symbolic operations are enabled (notably, the converse of any learned relation can be formed without additional training). Abstract semantic relations can be induced by bootstrapping from nonrelational inputs, thereby enabling relational generalization and analogical reasoning.
format Online
Article
Text
id pubmed-6410800
institution National Center for Biotechnology Information
language English
publishDate 2019
publisher National Academy of Sciences
record_format MEDLINE/PubMed
spelling pubmed-64108002019-03-13 Emergence of analogy from relation learning Lu, Hongjing Wu, Ying Nian Holyoak, Keith J. Proc Natl Acad Sci U S A Social Sciences By middle childhood, humans are able to learn abstract semantic relations (e.g., antonym, synonym, category membership) and use them to reason by analogy. A deep theoretical challenge is to show how such abstract relations can arise from nonrelational inputs, thereby providing key elements of a protosymbolic representation system. We have developed a computational model that exploits the potential synergy between deep learning from “big data” (to create semantic features for individual words) and supervised learning from “small data” (to create representations of semantic relations between words). Given as inputs labeled pairs of lexical representations extracted by deep learning, the model creates augmented representations by remapping features according to the rank of differences between values for the two words in each pair. These augmented representations aid in coping with the feature alignment problem (e.g., matching those features that make “love-hate” an antonym with the different features that make “rich-poor” an antonym). The model extracts weight distributions that are used to estimate the probabilities that new word pairs instantiate each relation, capturing the pattern of human typicality judgments for a broad range of abstract semantic relations. A measure of relational similarity can be derived and used to solve simple verbal analogies with human-level accuracy. Because each acquired relation has a modular representation, basic symbolic operations are enabled (notably, the converse of any learned relation can be formed without additional training). Abstract semantic relations can be induced by bootstrapping from nonrelational inputs, thereby enabling relational generalization and analogical reasoning. National Academy of Sciences 2019-03-05 2019-02-15 /pmc/articles/PMC6410800/ /pubmed/30770443 http://dx.doi.org/10.1073/pnas.1814779116 Text en Copyright © 2019 the Author(s). Published by PNAS. https://creativecommons.org/licenses/by-nc-nd/4.0/ This open access article is distributed under Creative Commons Attribution-NonCommercial-NoDerivatives License 4.0 (CC BY-NC-ND) (https://creativecommons.org/licenses/by-nc-nd/4.0/) .
spellingShingle Social Sciences
Lu, Hongjing
Wu, Ying Nian
Holyoak, Keith J.
Emergence of analogy from relation learning
title Emergence of analogy from relation learning
title_full Emergence of analogy from relation learning
title_fullStr Emergence of analogy from relation learning
title_full_unstemmed Emergence of analogy from relation learning
title_short Emergence of analogy from relation learning
title_sort emergence of analogy from relation learning
topic Social Sciences
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6410800/
https://www.ncbi.nlm.nih.gov/pubmed/30770443
http://dx.doi.org/10.1073/pnas.1814779116
work_keys_str_mv AT luhongjing emergenceofanalogyfromrelationlearning
AT wuyingnian emergenceofanalogyfromrelationlearning
AT holyoakkeithj emergenceofanalogyfromrelationlearning