Cargando…

Information-Theoretic Generalization Bounds for Meta-Learning and Applications

Meta-learning, or “learning to learn”, refers to techniques that infer an inductive bias from data corresponding to multiple related tasks with the goal of improving the sample efficiency for new, previously unobserved, tasks. A key performance measure for meta-learning is the meta-generalization ga...

Descripción completa

Detalles Bibliográficos
Autores principales: Jose, Sharu Theresa, Simeone, Osvaldo
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7835863/
https://www.ncbi.nlm.nih.gov/pubmed/33478002
http://dx.doi.org/10.3390/e23010126
_version_ 1783642626290876416
author Jose, Sharu Theresa
Simeone, Osvaldo
author_facet Jose, Sharu Theresa
Simeone, Osvaldo
author_sort Jose, Sharu Theresa
collection PubMed
description Meta-learning, or “learning to learn”, refers to techniques that infer an inductive bias from data corresponding to multiple related tasks with the goal of improving the sample efficiency for new, previously unobserved, tasks. A key performance measure for meta-learning is the meta-generalization gap, that is, the difference between the average loss measured on the meta-training data and on a new, randomly selected task. This paper presents novel information-theoretic upper bounds on the meta-generalization gap. Two broad classes of meta-learning algorithms are considered that use either separate within-task training and test sets, like model agnostic meta-learning (MAML), or joint within-task training and test sets, like reptile. Extending the existing work for conventional learning, an upper bound on the meta-generalization gap is derived for the former class that depends on the mutual information (MI) between the output of the meta-learning algorithm and its input meta-training data. For the latter, the derived bound includes an additional MI between the output of the per-task learning procedure and corresponding data set to capture within-task uncertainty. Tighter bounds are then developed for the two classes via novel individual task MI (ITMI) bounds. Applications of the derived bounds are finally discussed, including a broad class of noisy iterative algorithms for meta-learning.
format Online
Article
Text
id pubmed-7835863
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-78358632021-02-24 Information-Theoretic Generalization Bounds for Meta-Learning and Applications Jose, Sharu Theresa Simeone, Osvaldo Entropy (Basel) Article Meta-learning, or “learning to learn”, refers to techniques that infer an inductive bias from data corresponding to multiple related tasks with the goal of improving the sample efficiency for new, previously unobserved, tasks. A key performance measure for meta-learning is the meta-generalization gap, that is, the difference between the average loss measured on the meta-training data and on a new, randomly selected task. This paper presents novel information-theoretic upper bounds on the meta-generalization gap. Two broad classes of meta-learning algorithms are considered that use either separate within-task training and test sets, like model agnostic meta-learning (MAML), or joint within-task training and test sets, like reptile. Extending the existing work for conventional learning, an upper bound on the meta-generalization gap is derived for the former class that depends on the mutual information (MI) between the output of the meta-learning algorithm and its input meta-training data. For the latter, the derived bound includes an additional MI between the output of the per-task learning procedure and corresponding data set to capture within-task uncertainty. Tighter bounds are then developed for the two classes via novel individual task MI (ITMI) bounds. Applications of the derived bounds are finally discussed, including a broad class of noisy iterative algorithms for meta-learning. MDPI 2021-01-19 /pmc/articles/PMC7835863/ /pubmed/33478002 http://dx.doi.org/10.3390/e23010126 Text en © 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Jose, Sharu Theresa
Simeone, Osvaldo
Information-Theoretic Generalization Bounds for Meta-Learning and Applications
title Information-Theoretic Generalization Bounds for Meta-Learning and Applications
title_full Information-Theoretic Generalization Bounds for Meta-Learning and Applications
title_fullStr Information-Theoretic Generalization Bounds for Meta-Learning and Applications
title_full_unstemmed Information-Theoretic Generalization Bounds for Meta-Learning and Applications
title_short Information-Theoretic Generalization Bounds for Meta-Learning and Applications
title_sort information-theoretic generalization bounds for meta-learning and applications
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7835863/
https://www.ncbi.nlm.nih.gov/pubmed/33478002
http://dx.doi.org/10.3390/e23010126
work_keys_str_mv AT josesharutheresa informationtheoreticgeneralizationboundsformetalearningandapplications
AT simeoneosvaldo informationtheoreticgeneralizationboundsformetalearningandapplications