Cargando…

Towards a Unified Theory of Learning and Information

In this paper, we introduce the notion of “learning capacity” for algorithms that learn from data, which is analogous to the Shannon channel capacity for communication systems. We show how “learning capacity” bridges the gap between statistical learning theory and information theory, and we will use...

Descripción completa

Detalles Bibliográficos
Autor principal: Alabdulmohsin, Ibrahim
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7516920/
https://www.ncbi.nlm.nih.gov/pubmed/33286212
http://dx.doi.org/10.3390/e22040438
_version_ 1783587109728157696
author Alabdulmohsin, Ibrahim
author_facet Alabdulmohsin, Ibrahim
author_sort Alabdulmohsin, Ibrahim
collection PubMed
description In this paper, we introduce the notion of “learning capacity” for algorithms that learn from data, which is analogous to the Shannon channel capacity for communication systems. We show how “learning capacity” bridges the gap between statistical learning theory and information theory, and we will use it to derive generalization bounds for finite hypothesis spaces, differential privacy, and countable domains, among others. Moreover, we prove that under the Axiom of Choice, the existence of an empirical risk minimization (ERM) rule that has a vanishing learning capacity is equivalent to the assertion that the hypothesis space has a finite Vapnik–Chervonenkis (VC) dimension, thus establishing an equivalence relation between two of the most fundamental concepts in statistical learning theory and information theory. In addition, we show how the learning capacity of an algorithm provides important qualitative results, such as on the relation between generalization and algorithmic stability, information leakage, and data processing. Finally, we conclude by listing some open problems and suggesting future directions of research.
format Online
Article
Text
id pubmed-7516920
institution National Center for Biotechnology Information
language English
publishDate 2020
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-75169202020-11-09 Towards a Unified Theory of Learning and Information Alabdulmohsin, Ibrahim Entropy (Basel) Article In this paper, we introduce the notion of “learning capacity” for algorithms that learn from data, which is analogous to the Shannon channel capacity for communication systems. We show how “learning capacity” bridges the gap between statistical learning theory and information theory, and we will use it to derive generalization bounds for finite hypothesis spaces, differential privacy, and countable domains, among others. Moreover, we prove that under the Axiom of Choice, the existence of an empirical risk minimization (ERM) rule that has a vanishing learning capacity is equivalent to the assertion that the hypothesis space has a finite Vapnik–Chervonenkis (VC) dimension, thus establishing an equivalence relation between two of the most fundamental concepts in statistical learning theory and information theory. In addition, we show how the learning capacity of an algorithm provides important qualitative results, such as on the relation between generalization and algorithmic stability, information leakage, and data processing. Finally, we conclude by listing some open problems and suggesting future directions of research. MDPI 2020-04-13 /pmc/articles/PMC7516920/ /pubmed/33286212 http://dx.doi.org/10.3390/e22040438 Text en © 2020 by the author. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Alabdulmohsin, Ibrahim
Towards a Unified Theory of Learning and Information
title Towards a Unified Theory of Learning and Information
title_full Towards a Unified Theory of Learning and Information
title_fullStr Towards a Unified Theory of Learning and Information
title_full_unstemmed Towards a Unified Theory of Learning and Information
title_short Towards a Unified Theory of Learning and Information
title_sort towards a unified theory of learning and information
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7516920/
https://www.ncbi.nlm.nih.gov/pubmed/33286212
http://dx.doi.org/10.3390/e22040438
work_keys_str_mv AT alabdulmohsinibrahim towardsaunifiedtheoryoflearningandinformation