Cargando…
Self-Net: Lifelong Learning via Continual Self-Modeling
Learning a set of tasks over time, also known as continual learning (CL), is one of the most challenging problems in artificial intelligence. While recent approaches achieve some degree of CL in deep neural networks, they either (1) store a new network (or an equivalent number of parameters) for eac...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2020
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7861283/ https://www.ncbi.nlm.nih.gov/pubmed/33733138 http://dx.doi.org/10.3389/frai.2020.00019 |
_version_ | 1783647052579733504 |
---|---|
author | Mandivarapu, Jaya Krishna Camp, Blake Estrada, Rolando |
author_facet | Mandivarapu, Jaya Krishna Camp, Blake Estrada, Rolando |
author_sort | Mandivarapu, Jaya Krishna |
collection | PubMed |
description | Learning a set of tasks over time, also known as continual learning (CL), is one of the most challenging problems in artificial intelligence. While recent approaches achieve some degree of CL in deep neural networks, they either (1) store a new network (or an equivalent number of parameters) for each new task, (2) store training data from previous tasks, or (3) restrict the network's ability to learn new tasks. To address these issues, we propose a novel framework, Self-Net, that uses an autoencoder to learn a set of low-dimensional representations of the weights learned for different tasks. We demonstrate that these low-dimensional vectors can then be used to generate high-fidelity recollections of the original weights. Self-Net can incorporate new tasks over time with little retraining, minimal loss in performance for older tasks, and without storing prior training data. We show that our technique achieves over 10X storage compression in a continual fashion, and that it outperforms state-of-the-art approaches on numerous datasets, including continual versions of MNIST, CIFAR10, CIFAR100, Atari, and task-incremental CORe50. To the best of our knowledge, we are the first to use autoencoders to sequentially encode sets of network weights to enable continual learning. |
format | Online Article Text |
id | pubmed-7861283 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2020 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-78612832021-03-16 Self-Net: Lifelong Learning via Continual Self-Modeling Mandivarapu, Jaya Krishna Camp, Blake Estrada, Rolando Front Artif Intell Artificial Intelligence Learning a set of tasks over time, also known as continual learning (CL), is one of the most challenging problems in artificial intelligence. While recent approaches achieve some degree of CL in deep neural networks, they either (1) store a new network (or an equivalent number of parameters) for each new task, (2) store training data from previous tasks, or (3) restrict the network's ability to learn new tasks. To address these issues, we propose a novel framework, Self-Net, that uses an autoencoder to learn a set of low-dimensional representations of the weights learned for different tasks. We demonstrate that these low-dimensional vectors can then be used to generate high-fidelity recollections of the original weights. Self-Net can incorporate new tasks over time with little retraining, minimal loss in performance for older tasks, and without storing prior training data. We show that our technique achieves over 10X storage compression in a continual fashion, and that it outperforms state-of-the-art approaches on numerous datasets, including continual versions of MNIST, CIFAR10, CIFAR100, Atari, and task-incremental CORe50. To the best of our knowledge, we are the first to use autoencoders to sequentially encode sets of network weights to enable continual learning. Frontiers Media S.A. 2020-04-09 /pmc/articles/PMC7861283/ /pubmed/33733138 http://dx.doi.org/10.3389/frai.2020.00019 Text en Copyright © 2020 Mandivarapu, Camp and Estrada. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Artificial Intelligence Mandivarapu, Jaya Krishna Camp, Blake Estrada, Rolando Self-Net: Lifelong Learning via Continual Self-Modeling |
title | Self-Net: Lifelong Learning via Continual Self-Modeling |
title_full | Self-Net: Lifelong Learning via Continual Self-Modeling |
title_fullStr | Self-Net: Lifelong Learning via Continual Self-Modeling |
title_full_unstemmed | Self-Net: Lifelong Learning via Continual Self-Modeling |
title_short | Self-Net: Lifelong Learning via Continual Self-Modeling |
title_sort | self-net: lifelong learning via continual self-modeling |
topic | Artificial Intelligence |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7861283/ https://www.ncbi.nlm.nih.gov/pubmed/33733138 http://dx.doi.org/10.3389/frai.2020.00019 |
work_keys_str_mv | AT mandivarapujayakrishna selfnetlifelonglearningviacontinualselfmodeling AT campblake selfnetlifelonglearningviacontinualselfmodeling AT estradarolando selfnetlifelonglearningviacontinualselfmodeling |