Cargando…

Self-Net: Lifelong Learning via Continual Self-Modeling

Learning a set of tasks over time, also known as continual learning (CL), is one of the most challenging problems in artificial intelligence. While recent approaches achieve some degree of CL in deep neural networks, they either (1) store a new network (or an equivalent number of parameters) for eac...

Descripción completa

Detalles Bibliográficos
Autores principales: Mandivarapu, Jaya Krishna, Camp, Blake, Estrada, Rolando
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7861283/
https://www.ncbi.nlm.nih.gov/pubmed/33733138
http://dx.doi.org/10.3389/frai.2020.00019
_version_ 1783647052579733504
author Mandivarapu, Jaya Krishna
Camp, Blake
Estrada, Rolando
author_facet Mandivarapu, Jaya Krishna
Camp, Blake
Estrada, Rolando
author_sort Mandivarapu, Jaya Krishna
collection PubMed
description Learning a set of tasks over time, also known as continual learning (CL), is one of the most challenging problems in artificial intelligence. While recent approaches achieve some degree of CL in deep neural networks, they either (1) store a new network (or an equivalent number of parameters) for each new task, (2) store training data from previous tasks, or (3) restrict the network's ability to learn new tasks. To address these issues, we propose a novel framework, Self-Net, that uses an autoencoder to learn a set of low-dimensional representations of the weights learned for different tasks. We demonstrate that these low-dimensional vectors can then be used to generate high-fidelity recollections of the original weights. Self-Net can incorporate new tasks over time with little retraining, minimal loss in performance for older tasks, and without storing prior training data. We show that our technique achieves over 10X storage compression in a continual fashion, and that it outperforms state-of-the-art approaches on numerous datasets, including continual versions of MNIST, CIFAR10, CIFAR100, Atari, and task-incremental CORe50. To the best of our knowledge, we are the first to use autoencoders to sequentially encode sets of network weights to enable continual learning.
format Online
Article
Text
id pubmed-7861283
institution National Center for Biotechnology Information
language English
publishDate 2020
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-78612832021-03-16 Self-Net: Lifelong Learning via Continual Self-Modeling Mandivarapu, Jaya Krishna Camp, Blake Estrada, Rolando Front Artif Intell Artificial Intelligence Learning a set of tasks over time, also known as continual learning (CL), is one of the most challenging problems in artificial intelligence. While recent approaches achieve some degree of CL in deep neural networks, they either (1) store a new network (or an equivalent number of parameters) for each new task, (2) store training data from previous tasks, or (3) restrict the network's ability to learn new tasks. To address these issues, we propose a novel framework, Self-Net, that uses an autoencoder to learn a set of low-dimensional representations of the weights learned for different tasks. We demonstrate that these low-dimensional vectors can then be used to generate high-fidelity recollections of the original weights. Self-Net can incorporate new tasks over time with little retraining, minimal loss in performance for older tasks, and without storing prior training data. We show that our technique achieves over 10X storage compression in a continual fashion, and that it outperforms state-of-the-art approaches on numerous datasets, including continual versions of MNIST, CIFAR10, CIFAR100, Atari, and task-incremental CORe50. To the best of our knowledge, we are the first to use autoencoders to sequentially encode sets of network weights to enable continual learning. Frontiers Media S.A. 2020-04-09 /pmc/articles/PMC7861283/ /pubmed/33733138 http://dx.doi.org/10.3389/frai.2020.00019 Text en Copyright © 2020 Mandivarapu, Camp and Estrada. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Artificial Intelligence
Mandivarapu, Jaya Krishna
Camp, Blake
Estrada, Rolando
Self-Net: Lifelong Learning via Continual Self-Modeling
title Self-Net: Lifelong Learning via Continual Self-Modeling
title_full Self-Net: Lifelong Learning via Continual Self-Modeling
title_fullStr Self-Net: Lifelong Learning via Continual Self-Modeling
title_full_unstemmed Self-Net: Lifelong Learning via Continual Self-Modeling
title_short Self-Net: Lifelong Learning via Continual Self-Modeling
title_sort self-net: lifelong learning via continual self-modeling
topic Artificial Intelligence
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7861283/
https://www.ncbi.nlm.nih.gov/pubmed/33733138
http://dx.doi.org/10.3389/frai.2020.00019
work_keys_str_mv AT mandivarapujayakrishna selfnetlifelonglearningviacontinualselfmodeling
AT campblake selfnetlifelonglearningviacontinualselfmodeling
AT estradarolando selfnetlifelonglearningviacontinualselfmodeling