Cargando…
Information Flows of Diverse Autoencoders
Deep learning methods have had outstanding performances in various fields. A fundamental query is why they are so effective. Information theory provides a potential answer by interpreting the learning process as the information transmission and compression of data. The information flows can be visua...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8303402/ https://www.ncbi.nlm.nih.gov/pubmed/34356403 http://dx.doi.org/10.3390/e23070862 |
_version_ | 1783727078335578112 |
---|---|
author | Lee, Sungyeop Jo, Junghyo |
author_facet | Lee, Sungyeop Jo, Junghyo |
author_sort | Lee, Sungyeop |
collection | PubMed |
description | Deep learning methods have had outstanding performances in various fields. A fundamental query is why they are so effective. Information theory provides a potential answer by interpreting the learning process as the information transmission and compression of data. The information flows can be visualized on the information plane of the mutual information among the input, hidden, and output layers. In this study, we examine how the information flows are shaped by the network parameters, such as depth, sparsity, weight constraints, and hidden representations. Here, we adopt autoencoders as models of deep learning, because (i) they have clear guidelines for their information flows, and (ii) they have various species, such as vanilla, sparse, tied, variational, and label autoencoders. We measured their information flows using Rényi’s matrix-based [Formula: see text]-order entropy functional. As learning progresses, they show a typical fitting phase where the amounts of input-to-hidden and hidden-to-output mutual information both increase. In the last stage of learning, however, some autoencoders show a simplifying phase, previously called the “compression phase”, where input-to-hidden mutual information diminishes. In particular, the sparsity regularization of hidden activities amplifies the simplifying phase. However, tied, variational, and label autoencoders do not have a simplifying phase. Nevertheless, all autoencoders have similar reconstruction errors for training and test data. Thus, the simplifying phase does not seem to be necessary for the generalization of learning. |
format | Online Article Text |
id | pubmed-8303402 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-83034022021-07-25 Information Flows of Diverse Autoencoders Lee, Sungyeop Jo, Junghyo Entropy (Basel) Article Deep learning methods have had outstanding performances in various fields. A fundamental query is why they are so effective. Information theory provides a potential answer by interpreting the learning process as the information transmission and compression of data. The information flows can be visualized on the information plane of the mutual information among the input, hidden, and output layers. In this study, we examine how the information flows are shaped by the network parameters, such as depth, sparsity, weight constraints, and hidden representations. Here, we adopt autoencoders as models of deep learning, because (i) they have clear guidelines for their information flows, and (ii) they have various species, such as vanilla, sparse, tied, variational, and label autoencoders. We measured their information flows using Rényi’s matrix-based [Formula: see text]-order entropy functional. As learning progresses, they show a typical fitting phase where the amounts of input-to-hidden and hidden-to-output mutual information both increase. In the last stage of learning, however, some autoencoders show a simplifying phase, previously called the “compression phase”, where input-to-hidden mutual information diminishes. In particular, the sparsity regularization of hidden activities amplifies the simplifying phase. However, tied, variational, and label autoencoders do not have a simplifying phase. Nevertheless, all autoencoders have similar reconstruction errors for training and test data. Thus, the simplifying phase does not seem to be necessary for the generalization of learning. MDPI 2021-07-05 /pmc/articles/PMC8303402/ /pubmed/34356403 http://dx.doi.org/10.3390/e23070862 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Lee, Sungyeop Jo, Junghyo Information Flows of Diverse Autoencoders |
title | Information Flows of Diverse Autoencoders |
title_full | Information Flows of Diverse Autoencoders |
title_fullStr | Information Flows of Diverse Autoencoders |
title_full_unstemmed | Information Flows of Diverse Autoencoders |
title_short | Information Flows of Diverse Autoencoders |
title_sort | information flows of diverse autoencoders |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8303402/ https://www.ncbi.nlm.nih.gov/pubmed/34356403 http://dx.doi.org/10.3390/e23070862 |
work_keys_str_mv | AT leesungyeop informationflowsofdiverseautoencoders AT jojunghyo informationflowsofdiverseautoencoders |