Cargando…
Meta-Strategy for Learning Tuning Parameters with Guarantees
Online learning methods, similar to the online gradient algorithm (OGA) and exponentially weighted aggregation (EWA), often depend on tuning parameters that are difficult to set in practice. We consider an online meta-learning scenario, and we propose a meta-strategy to learn these parameters from p...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8534330/ https://www.ncbi.nlm.nih.gov/pubmed/34681980 http://dx.doi.org/10.3390/e23101257 |
_version_ | 1784587526855458816 |
---|---|
author | Meunier, Dimitri Alquier, Pierre |
author_facet | Meunier, Dimitri Alquier, Pierre |
author_sort | Meunier, Dimitri |
collection | PubMed |
description | Online learning methods, similar to the online gradient algorithm (OGA) and exponentially weighted aggregation (EWA), often depend on tuning parameters that are difficult to set in practice. We consider an online meta-learning scenario, and we propose a meta-strategy to learn these parameters from past tasks. Our strategy is based on the minimization of a regret bound. It allows us to learn the initialization and the step size in OGA with guarantees. It also allows us to learn the prior or the learning rate in EWA. We provide a regret analysis of the strategy. It allows to identify settings where meta-learning indeed improves on learning each task in isolation. |
format | Online Article Text |
id | pubmed-8534330 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-85343302021-10-23 Meta-Strategy for Learning Tuning Parameters with Guarantees Meunier, Dimitri Alquier, Pierre Entropy (Basel) Article Online learning methods, similar to the online gradient algorithm (OGA) and exponentially weighted aggregation (EWA), often depend on tuning parameters that are difficult to set in practice. We consider an online meta-learning scenario, and we propose a meta-strategy to learn these parameters from past tasks. Our strategy is based on the minimization of a regret bound. It allows us to learn the initialization and the step size in OGA with guarantees. It also allows us to learn the prior or the learning rate in EWA. We provide a regret analysis of the strategy. It allows to identify settings where meta-learning indeed improves on learning each task in isolation. MDPI 2021-09-27 /pmc/articles/PMC8534330/ /pubmed/34681980 http://dx.doi.org/10.3390/e23101257 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Meunier, Dimitri Alquier, Pierre Meta-Strategy for Learning Tuning Parameters with Guarantees |
title | Meta-Strategy for Learning Tuning Parameters with Guarantees |
title_full | Meta-Strategy for Learning Tuning Parameters with Guarantees |
title_fullStr | Meta-Strategy for Learning Tuning Parameters with Guarantees |
title_full_unstemmed | Meta-Strategy for Learning Tuning Parameters with Guarantees |
title_short | Meta-Strategy for Learning Tuning Parameters with Guarantees |
title_sort | meta-strategy for learning tuning parameters with guarantees |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8534330/ https://www.ncbi.nlm.nih.gov/pubmed/34681980 http://dx.doi.org/10.3390/e23101257 |
work_keys_str_mv | AT meunierdimitri metastrategyforlearningtuningparameterswithguarantees AT alquierpierre metastrategyforlearningtuningparameterswithguarantees |