Cargando…

Still No Free Lunches: The Price to Pay for Tighter PAC-Bayes Bounds

“No free lunch” results state the impossibility of obtaining meaningful bounds on the error of a learning algorithm without prior assumptions and modelling, which is more or less realistic for a given problem. Some models are “expensive” (strong assumptions, such as sub-Gaussian tails), others are “...

Descripción completa

Detalles Bibliográficos
Autores principales: Guedj, Benjamin, Pujol, Louis
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8619508/
https://www.ncbi.nlm.nih.gov/pubmed/34828227
http://dx.doi.org/10.3390/e23111529
_version_ 1784605009637277696
author Guedj, Benjamin
Pujol, Louis
author_facet Guedj, Benjamin
Pujol, Louis
author_sort Guedj, Benjamin
collection PubMed
description “No free lunch” results state the impossibility of obtaining meaningful bounds on the error of a learning algorithm without prior assumptions and modelling, which is more or less realistic for a given problem. Some models are “expensive” (strong assumptions, such as sub-Gaussian tails), others are “cheap” (simply finite variance). As it is well known, the more you pay, the more you get: in other words, the most expensive models yield the more interesting bounds. Recent advances in robust statistics have investigated procedures to obtain tight bounds while keeping the cost of assumptions minimal. The present paper explores and exhibits what the limits are for obtaining tight probably approximately correct (PAC)-Bayes bounds in a robust setting for cheap models.
format Online
Article
Text
id pubmed-8619508
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-86195082021-11-27 Still No Free Lunches: The Price to Pay for Tighter PAC-Bayes Bounds Guedj, Benjamin Pujol, Louis Entropy (Basel) Article “No free lunch” results state the impossibility of obtaining meaningful bounds on the error of a learning algorithm without prior assumptions and modelling, which is more or less realistic for a given problem. Some models are “expensive” (strong assumptions, such as sub-Gaussian tails), others are “cheap” (simply finite variance). As it is well known, the more you pay, the more you get: in other words, the most expensive models yield the more interesting bounds. Recent advances in robust statistics have investigated procedures to obtain tight bounds while keeping the cost of assumptions minimal. The present paper explores and exhibits what the limits are for obtaining tight probably approximately correct (PAC)-Bayes bounds in a robust setting for cheap models. MDPI 2021-11-18 /pmc/articles/PMC8619508/ /pubmed/34828227 http://dx.doi.org/10.3390/e23111529 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Guedj, Benjamin
Pujol, Louis
Still No Free Lunches: The Price to Pay for Tighter PAC-Bayes Bounds
title Still No Free Lunches: The Price to Pay for Tighter PAC-Bayes Bounds
title_full Still No Free Lunches: The Price to Pay for Tighter PAC-Bayes Bounds
title_fullStr Still No Free Lunches: The Price to Pay for Tighter PAC-Bayes Bounds
title_full_unstemmed Still No Free Lunches: The Price to Pay for Tighter PAC-Bayes Bounds
title_short Still No Free Lunches: The Price to Pay for Tighter PAC-Bayes Bounds
title_sort still no free lunches: the price to pay for tighter pac-bayes bounds
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8619508/
https://www.ncbi.nlm.nih.gov/pubmed/34828227
http://dx.doi.org/10.3390/e23111529
work_keys_str_mv AT guedjbenjamin stillnofreelunchesthepricetopayfortighterpacbayesbounds
AT pujollouis stillnofreelunchesthepricetopayfortighterpacbayesbounds