Cargando…
Sleep-like unsupervised replay reduces catastrophic forgetting in artificial neural networks
Artificial neural networks are known to suffer from catastrophic forgetting: when learning multiple tasks sequentially, they perform well on the most recent task at the expense of previously learned tasks. In the brain, sleep is known to play an important role in incremental learning by replaying re...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Nature Publishing Group UK
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9755223/ https://www.ncbi.nlm.nih.gov/pubmed/36522325 http://dx.doi.org/10.1038/s41467-022-34938-7 |
_version_ | 1784851382560358400 |
---|---|
author | Tadros, Timothy Krishnan, Giri P. Ramyaa, Ramyaa Bazhenov, Maxim |
author_facet | Tadros, Timothy Krishnan, Giri P. Ramyaa, Ramyaa Bazhenov, Maxim |
author_sort | Tadros, Timothy |
collection | PubMed |
description | Artificial neural networks are known to suffer from catastrophic forgetting: when learning multiple tasks sequentially, they perform well on the most recent task at the expense of previously learned tasks. In the brain, sleep is known to play an important role in incremental learning by replaying recent and old conflicting memory traces. Here we tested the hypothesis that implementing a sleep-like phase in artificial neural networks can protect old memories during new training and alleviate catastrophic forgetting. Sleep was implemented as off-line training with local unsupervised Hebbian plasticity rules and noisy input. In an incremental learning framework, sleep was able to recover old tasks that were otherwise forgotten. Previously learned memories were replayed spontaneously during sleep, forming unique representations for each class of inputs. Representational sparseness and neuronal activity corresponding to the old tasks increased while new task related activity decreased. The study suggests that spontaneous replay simulating sleep-like dynamics can alleviate catastrophic forgetting in artificial neural networks. |
format | Online Article Text |
id | pubmed-9755223 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Nature Publishing Group UK |
record_format | MEDLINE/PubMed |
spelling | pubmed-97552232022-12-17 Sleep-like unsupervised replay reduces catastrophic forgetting in artificial neural networks Tadros, Timothy Krishnan, Giri P. Ramyaa, Ramyaa Bazhenov, Maxim Nat Commun Article Artificial neural networks are known to suffer from catastrophic forgetting: when learning multiple tasks sequentially, they perform well on the most recent task at the expense of previously learned tasks. In the brain, sleep is known to play an important role in incremental learning by replaying recent and old conflicting memory traces. Here we tested the hypothesis that implementing a sleep-like phase in artificial neural networks can protect old memories during new training and alleviate catastrophic forgetting. Sleep was implemented as off-line training with local unsupervised Hebbian plasticity rules and noisy input. In an incremental learning framework, sleep was able to recover old tasks that were otherwise forgotten. Previously learned memories were replayed spontaneously during sleep, forming unique representations for each class of inputs. Representational sparseness and neuronal activity corresponding to the old tasks increased while new task related activity decreased. The study suggests that spontaneous replay simulating sleep-like dynamics can alleviate catastrophic forgetting in artificial neural networks. Nature Publishing Group UK 2022-12-15 /pmc/articles/PMC9755223/ /pubmed/36522325 http://dx.doi.org/10.1038/s41467-022-34938-7 Text en © The Author(s) 2022 https://creativecommons.org/licenses/by/4.0/Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) . |
spellingShingle | Article Tadros, Timothy Krishnan, Giri P. Ramyaa, Ramyaa Bazhenov, Maxim Sleep-like unsupervised replay reduces catastrophic forgetting in artificial neural networks |
title | Sleep-like unsupervised replay reduces catastrophic forgetting in artificial neural networks |
title_full | Sleep-like unsupervised replay reduces catastrophic forgetting in artificial neural networks |
title_fullStr | Sleep-like unsupervised replay reduces catastrophic forgetting in artificial neural networks |
title_full_unstemmed | Sleep-like unsupervised replay reduces catastrophic forgetting in artificial neural networks |
title_short | Sleep-like unsupervised replay reduces catastrophic forgetting in artificial neural networks |
title_sort | sleep-like unsupervised replay reduces catastrophic forgetting in artificial neural networks |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9755223/ https://www.ncbi.nlm.nih.gov/pubmed/36522325 http://dx.doi.org/10.1038/s41467-022-34938-7 |
work_keys_str_mv | AT tadrostimothy sleeplikeunsupervisedreplayreducescatastrophicforgettinginartificialneuralnetworks AT krishnangirip sleeplikeunsupervisedreplayreducescatastrophicforgettinginartificialneuralnetworks AT ramyaaramyaa sleeplikeunsupervisedreplayreducescatastrophicforgettinginartificialneuralnetworks AT bazhenovmaxim sleeplikeunsupervisedreplayreducescatastrophicforgettinginartificialneuralnetworks |