Cargando…

Rapid runtime learning by curating small datasets of high-quality items obtained from memory

We propose the “runtime learning” hypothesis which states that people quickly learn to perform unfamiliar tasks as the tasks arise by using task-relevant instances of concepts stored in memory during mental training. To make learning rapid, the hypothesis claims that only a few class instances are u...

Descripción completa

Detalles Bibliográficos
Autores principales: German, Joseph Scott, Cui, Guofeng, Xu, Chenliang, Jacobs, Robert A.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Public Library of Science 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10578607/
https://www.ncbi.nlm.nih.gov/pubmed/37792896
http://dx.doi.org/10.1371/journal.pcbi.1011445
_version_ 1785121553544904704
author German, Joseph Scott
Cui, Guofeng
Xu, Chenliang
Jacobs, Robert A.
author_facet German, Joseph Scott
Cui, Guofeng
Xu, Chenliang
Jacobs, Robert A.
author_sort German, Joseph Scott
collection PubMed
description We propose the “runtime learning” hypothesis which states that people quickly learn to perform unfamiliar tasks as the tasks arise by using task-relevant instances of concepts stored in memory during mental training. To make learning rapid, the hypothesis claims that only a few class instances are used, but these instances are especially valuable for training. The paper motivates the hypothesis by describing related ideas from the cognitive science and machine learning literatures. Using computer simulation, we show that deep neural networks (DNNs) can learn effectively from small, curated training sets, and that valuable training items tend to lie toward the centers of data item clusters in an abstract feature space. In a series of three behavioral experiments, we show that people can also learn effectively from small, curated training sets. Critically, we find that participant reaction times and fitted drift rates are best accounted for by the confidences of DNNs trained on small datasets of highly valuable items. We conclude that the runtime learning hypothesis is a novel conjecture about the relationship between learning and memory with the potential for explaining a wide variety of cognitive phenomena.
format Online
Article
Text
id pubmed-10578607
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-105786072023-10-17 Rapid runtime learning by curating small datasets of high-quality items obtained from memory German, Joseph Scott Cui, Guofeng Xu, Chenliang Jacobs, Robert A. PLoS Comput Biol Research Article We propose the “runtime learning” hypothesis which states that people quickly learn to perform unfamiliar tasks as the tasks arise by using task-relevant instances of concepts stored in memory during mental training. To make learning rapid, the hypothesis claims that only a few class instances are used, but these instances are especially valuable for training. The paper motivates the hypothesis by describing related ideas from the cognitive science and machine learning literatures. Using computer simulation, we show that deep neural networks (DNNs) can learn effectively from small, curated training sets, and that valuable training items tend to lie toward the centers of data item clusters in an abstract feature space. In a series of three behavioral experiments, we show that people can also learn effectively from small, curated training sets. Critically, we find that participant reaction times and fitted drift rates are best accounted for by the confidences of DNNs trained on small datasets of highly valuable items. We conclude that the runtime learning hypothesis is a novel conjecture about the relationship between learning and memory with the potential for explaining a wide variety of cognitive phenomena. Public Library of Science 2023-10-04 /pmc/articles/PMC10578607/ /pubmed/37792896 http://dx.doi.org/10.1371/journal.pcbi.1011445 Text en © 2023 German et al https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
spellingShingle Research Article
German, Joseph Scott
Cui, Guofeng
Xu, Chenliang
Jacobs, Robert A.
Rapid runtime learning by curating small datasets of high-quality items obtained from memory
title Rapid runtime learning by curating small datasets of high-quality items obtained from memory
title_full Rapid runtime learning by curating small datasets of high-quality items obtained from memory
title_fullStr Rapid runtime learning by curating small datasets of high-quality items obtained from memory
title_full_unstemmed Rapid runtime learning by curating small datasets of high-quality items obtained from memory
title_short Rapid runtime learning by curating small datasets of high-quality items obtained from memory
title_sort rapid runtime learning by curating small datasets of high-quality items obtained from memory
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10578607/
https://www.ncbi.nlm.nih.gov/pubmed/37792896
http://dx.doi.org/10.1371/journal.pcbi.1011445
work_keys_str_mv AT germanjosephscott rapidruntimelearningbycuratingsmalldatasetsofhighqualityitemsobtainedfrommemory
AT cuiguofeng rapidruntimelearningbycuratingsmalldatasetsofhighqualityitemsobtainedfrommemory
AT xuchenliang rapidruntimelearningbycuratingsmalldatasetsofhighqualityitemsobtainedfrommemory
AT jacobsroberta rapidruntimelearningbycuratingsmalldatasetsofhighqualityitemsobtainedfrommemory