Cargando…

On Sequential Bayesian Inference for Continual Learning

Sequential Bayesian inference can be used for continual learning to prevent catastrophic forgetting of past tasks and provide an informative prior when learning new tasks. We revisit sequential Bayesian inference and assess whether using the previous task’s posterior as a prior for a new task can pr...

Descripción completa

Detalles Bibliográficos
Autores principales: Kessler, Samuel, Cobb, Adam, Rudner, Tim G. J., Zohren, Stefan, Roberts, Stephen J.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10297370/
https://www.ncbi.nlm.nih.gov/pubmed/37372228
http://dx.doi.org/10.3390/e25060884
_version_ 1785063867821326336
author Kessler, Samuel
Cobb, Adam
Rudner, Tim G. J.
Zohren, Stefan
Roberts, Stephen J.
author_facet Kessler, Samuel
Cobb, Adam
Rudner, Tim G. J.
Zohren, Stefan
Roberts, Stephen J.
author_sort Kessler, Samuel
collection PubMed
description Sequential Bayesian inference can be used for continual learning to prevent catastrophic forgetting of past tasks and provide an informative prior when learning new tasks. We revisit sequential Bayesian inference and assess whether using the previous task’s posterior as a prior for a new task can prevent catastrophic forgetting in Bayesian neural networks. Our first contribution is to perform sequential Bayesian inference using Hamiltonian Monte Carlo. We propagate the posterior as a prior for new tasks by approximating the posterior via fitting a density estimator on Hamiltonian Monte Carlo samples. We find that this approach fails to prevent catastrophic forgetting, demonstrating the difficulty in performing sequential Bayesian inference in neural networks. From there, we study simple analytical examples of sequential Bayesian inference and CL and highlight the issue of model misspecification, which can lead to sub-optimal continual learning performance despite exact inference. Furthermore, we discuss how task data imbalances can cause forgetting. From these limitations, we argue that we need probabilistic models of the continual learning generative process rather than relying on sequential Bayesian inference over Bayesian neural network weights. Our final contribution is to propose a simple baseline called Prototypical Bayesian Continual Learning, which is competitive with the best performing Bayesian continual learning methods on class incremental continual learning computer vision benchmarks.
format Online
Article
Text
id pubmed-10297370
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-102973702023-06-28 On Sequential Bayesian Inference for Continual Learning Kessler, Samuel Cobb, Adam Rudner, Tim G. J. Zohren, Stefan Roberts, Stephen J. Entropy (Basel) Article Sequential Bayesian inference can be used for continual learning to prevent catastrophic forgetting of past tasks and provide an informative prior when learning new tasks. We revisit sequential Bayesian inference and assess whether using the previous task’s posterior as a prior for a new task can prevent catastrophic forgetting in Bayesian neural networks. Our first contribution is to perform sequential Bayesian inference using Hamiltonian Monte Carlo. We propagate the posterior as a prior for new tasks by approximating the posterior via fitting a density estimator on Hamiltonian Monte Carlo samples. We find that this approach fails to prevent catastrophic forgetting, demonstrating the difficulty in performing sequential Bayesian inference in neural networks. From there, we study simple analytical examples of sequential Bayesian inference and CL and highlight the issue of model misspecification, which can lead to sub-optimal continual learning performance despite exact inference. Furthermore, we discuss how task data imbalances can cause forgetting. From these limitations, we argue that we need probabilistic models of the continual learning generative process rather than relying on sequential Bayesian inference over Bayesian neural network weights. Our final contribution is to propose a simple baseline called Prototypical Bayesian Continual Learning, which is competitive with the best performing Bayesian continual learning methods on class incremental continual learning computer vision benchmarks. MDPI 2023-05-31 /pmc/articles/PMC10297370/ /pubmed/37372228 http://dx.doi.org/10.3390/e25060884 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Kessler, Samuel
Cobb, Adam
Rudner, Tim G. J.
Zohren, Stefan
Roberts, Stephen J.
On Sequential Bayesian Inference for Continual Learning
title On Sequential Bayesian Inference for Continual Learning
title_full On Sequential Bayesian Inference for Continual Learning
title_fullStr On Sequential Bayesian Inference for Continual Learning
title_full_unstemmed On Sequential Bayesian Inference for Continual Learning
title_short On Sequential Bayesian Inference for Continual Learning
title_sort on sequential bayesian inference for continual learning
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10297370/
https://www.ncbi.nlm.nih.gov/pubmed/37372228
http://dx.doi.org/10.3390/e25060884
work_keys_str_mv AT kesslersamuel onsequentialbayesianinferenceforcontinuallearning
AT cobbadam onsequentialbayesianinferenceforcontinuallearning
AT rudnertimgj onsequentialbayesianinferenceforcontinuallearning
AT zohrenstefan onsequentialbayesianinferenceforcontinuallearning
AT robertsstephenj onsequentialbayesianinferenceforcontinuallearning