Cargando…

Meta-Strategy for Learning Tuning Parameters with Guarantees

Online learning methods, similar to the online gradient algorithm (OGA) and exponentially weighted aggregation (EWA), often depend on tuning parameters that are difficult to set in practice. We consider an online meta-learning scenario, and we propose a meta-strategy to learn these parameters from p...

Descripción completa

Detalles Bibliográficos
Autores principales: Meunier, Dimitri, Alquier, Pierre
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8534330/
https://www.ncbi.nlm.nih.gov/pubmed/34681980
http://dx.doi.org/10.3390/e23101257
_version_ 1784587526855458816
author Meunier, Dimitri
Alquier, Pierre
author_facet Meunier, Dimitri
Alquier, Pierre
author_sort Meunier, Dimitri
collection PubMed
description Online learning methods, similar to the online gradient algorithm (OGA) and exponentially weighted aggregation (EWA), often depend on tuning parameters that are difficult to set in practice. We consider an online meta-learning scenario, and we propose a meta-strategy to learn these parameters from past tasks. Our strategy is based on the minimization of a regret bound. It allows us to learn the initialization and the step size in OGA with guarantees. It also allows us to learn the prior or the learning rate in EWA. We provide a regret analysis of the strategy. It allows to identify settings where meta-learning indeed improves on learning each task in isolation.
format Online
Article
Text
id pubmed-8534330
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-85343302021-10-23 Meta-Strategy for Learning Tuning Parameters with Guarantees Meunier, Dimitri Alquier, Pierre Entropy (Basel) Article Online learning methods, similar to the online gradient algorithm (OGA) and exponentially weighted aggregation (EWA), often depend on tuning parameters that are difficult to set in practice. We consider an online meta-learning scenario, and we propose a meta-strategy to learn these parameters from past tasks. Our strategy is based on the minimization of a regret bound. It allows us to learn the initialization and the step size in OGA with guarantees. It also allows us to learn the prior or the learning rate in EWA. We provide a regret analysis of the strategy. It allows to identify settings where meta-learning indeed improves on learning each task in isolation. MDPI 2021-09-27 /pmc/articles/PMC8534330/ /pubmed/34681980 http://dx.doi.org/10.3390/e23101257 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Meunier, Dimitri
Alquier, Pierre
Meta-Strategy for Learning Tuning Parameters with Guarantees
title Meta-Strategy for Learning Tuning Parameters with Guarantees
title_full Meta-Strategy for Learning Tuning Parameters with Guarantees
title_fullStr Meta-Strategy for Learning Tuning Parameters with Guarantees
title_full_unstemmed Meta-Strategy for Learning Tuning Parameters with Guarantees
title_short Meta-Strategy for Learning Tuning Parameters with Guarantees
title_sort meta-strategy for learning tuning parameters with guarantees
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8534330/
https://www.ncbi.nlm.nih.gov/pubmed/34681980
http://dx.doi.org/10.3390/e23101257
work_keys_str_mv AT meunierdimitri metastrategyforlearningtuningparameterswithguarantees
AT alquierpierre metastrategyforlearningtuningparameterswithguarantees