Cargando…

Learning in deep neural networks and brains with similarity-weighted interleaved learning

Understanding how the brain learns throughout a lifetime remains a long-standing challenge. In artificial neural networks (ANNs), incorporating novel information too rapidly results in catastrophic interference, i.e., abrupt loss of previously acquired knowledge. Complementary Learning Systems Theor...

Descripción completa

Detalles Bibliográficos
Autores principales: Saxena, Rajat, Shobe, Justin L., McNaughton, Bruce L.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: National Academy of Sciences 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9271163/
https://www.ncbi.nlm.nih.gov/pubmed/35759669
http://dx.doi.org/10.1073/pnas.2115229119
_version_ 1784744619246878720
author Saxena, Rajat
Shobe, Justin L.
McNaughton, Bruce L.
author_facet Saxena, Rajat
Shobe, Justin L.
McNaughton, Bruce L.
author_sort Saxena, Rajat
collection PubMed
description Understanding how the brain learns throughout a lifetime remains a long-standing challenge. In artificial neural networks (ANNs), incorporating novel information too rapidly results in catastrophic interference, i.e., abrupt loss of previously acquired knowledge. Complementary Learning Systems Theory (CLST) suggests that new memories can be gradually integrated into the neocortex by interleaving new memories with existing knowledge. This approach, however, has been assumed to require interleaving all existing knowledge every time something new is learned, which is implausible because it is time-consuming and requires a large amount of data. We show that deep, nonlinear ANNs can learn new information by interleaving only a subset of old items that share substantial representational similarity with the new information. By using such similarity-weighted interleaved learning (SWIL), ANNs can learn new information rapidly with a similar accuracy level and minimal interference, while using a much smaller number of old items presented per epoch (fast and data-efficient). SWIL is shown to work with various standard classification datasets (Fashion-MNIST, CIFAR10, and CIFAR100), deep neural network architectures, and in sequential learning frameworks. We show that data efficiency and speedup in learning new items are increased roughly proportionally to the number of nonoverlapping classes stored in the network, which implies an enormous possible speedup in human brains, which encode a high number of separate categories. Finally, we propose a theoretical model of how SWIL might be implemented in the brain.
format Online
Article
Text
id pubmed-9271163
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher National Academy of Sciences
record_format MEDLINE/PubMed
spelling pubmed-92711632022-07-11 Learning in deep neural networks and brains with similarity-weighted interleaved learning Saxena, Rajat Shobe, Justin L. McNaughton, Bruce L. Proc Natl Acad Sci U S A Biological Sciences Understanding how the brain learns throughout a lifetime remains a long-standing challenge. In artificial neural networks (ANNs), incorporating novel information too rapidly results in catastrophic interference, i.e., abrupt loss of previously acquired knowledge. Complementary Learning Systems Theory (CLST) suggests that new memories can be gradually integrated into the neocortex by interleaving new memories with existing knowledge. This approach, however, has been assumed to require interleaving all existing knowledge every time something new is learned, which is implausible because it is time-consuming and requires a large amount of data. We show that deep, nonlinear ANNs can learn new information by interleaving only a subset of old items that share substantial representational similarity with the new information. By using such similarity-weighted interleaved learning (SWIL), ANNs can learn new information rapidly with a similar accuracy level and minimal interference, while using a much smaller number of old items presented per epoch (fast and data-efficient). SWIL is shown to work with various standard classification datasets (Fashion-MNIST, CIFAR10, and CIFAR100), deep neural network architectures, and in sequential learning frameworks. We show that data efficiency and speedup in learning new items are increased roughly proportionally to the number of nonoverlapping classes stored in the network, which implies an enormous possible speedup in human brains, which encode a high number of separate categories. Finally, we propose a theoretical model of how SWIL might be implemented in the brain. National Academy of Sciences 2022-06-27 2022-07-05 /pmc/articles/PMC9271163/ /pubmed/35759669 http://dx.doi.org/10.1073/pnas.2115229119 Text en Copyright © 2022 the Author(s). Published by PNAS https://creativecommons.org/licenses/by-nc-nd/4.0/This open access article is distributed under Creative Commons Attribution-NonCommercial-NoDerivatives License 4.0 (CC BY-NC-ND) (https://creativecommons.org/licenses/by-nc-nd/4.0/) .
spellingShingle Biological Sciences
Saxena, Rajat
Shobe, Justin L.
McNaughton, Bruce L.
Learning in deep neural networks and brains with similarity-weighted interleaved learning
title Learning in deep neural networks and brains with similarity-weighted interleaved learning
title_full Learning in deep neural networks and brains with similarity-weighted interleaved learning
title_fullStr Learning in deep neural networks and brains with similarity-weighted interleaved learning
title_full_unstemmed Learning in deep neural networks and brains with similarity-weighted interleaved learning
title_short Learning in deep neural networks and brains with similarity-weighted interleaved learning
title_sort learning in deep neural networks and brains with similarity-weighted interleaved learning
topic Biological Sciences
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9271163/
https://www.ncbi.nlm.nih.gov/pubmed/35759669
http://dx.doi.org/10.1073/pnas.2115229119
work_keys_str_mv AT saxenarajat learningindeepneuralnetworksandbrainswithsimilarityweightedinterleavedlearning
AT shobejustinl learningindeepneuralnetworksandbrainswithsimilarityweightedinterleavedlearning
AT mcnaughtonbrucel learningindeepneuralnetworksandbrainswithsimilarityweightedinterleavedlearning