Cargando…

Information Flows of Diverse Autoencoders

Deep learning methods have had outstanding performances in various fields. A fundamental query is why they are so effective. Information theory provides a potential answer by interpreting the learning process as the information transmission and compression of data. The information flows can be visua...

Descripción completa

Detalles Bibliográficos
Autores principales: Lee, Sungyeop, Jo, Junghyo
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8303402/
https://www.ncbi.nlm.nih.gov/pubmed/34356403
http://dx.doi.org/10.3390/e23070862
_version_ 1783727078335578112
author Lee, Sungyeop
Jo, Junghyo
author_facet Lee, Sungyeop
Jo, Junghyo
author_sort Lee, Sungyeop
collection PubMed
description Deep learning methods have had outstanding performances in various fields. A fundamental query is why they are so effective. Information theory provides a potential answer by interpreting the learning process as the information transmission and compression of data. The information flows can be visualized on the information plane of the mutual information among the input, hidden, and output layers. In this study, we examine how the information flows are shaped by the network parameters, such as depth, sparsity, weight constraints, and hidden representations. Here, we adopt autoencoders as models of deep learning, because (i) they have clear guidelines for their information flows, and (ii) they have various species, such as vanilla, sparse, tied, variational, and label autoencoders. We measured their information flows using Rényi’s matrix-based [Formula: see text]-order entropy functional. As learning progresses, they show a typical fitting phase where the amounts of input-to-hidden and hidden-to-output mutual information both increase. In the last stage of learning, however, some autoencoders show a simplifying phase, previously called the “compression phase”, where input-to-hidden mutual information diminishes. In particular, the sparsity regularization of hidden activities amplifies the simplifying phase. However, tied, variational, and label autoencoders do not have a simplifying phase. Nevertheless, all autoencoders have similar reconstruction errors for training and test data. Thus, the simplifying phase does not seem to be necessary for the generalization of learning.
format Online
Article
Text
id pubmed-8303402
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-83034022021-07-25 Information Flows of Diverse Autoencoders Lee, Sungyeop Jo, Junghyo Entropy (Basel) Article Deep learning methods have had outstanding performances in various fields. A fundamental query is why they are so effective. Information theory provides a potential answer by interpreting the learning process as the information transmission and compression of data. The information flows can be visualized on the information plane of the mutual information among the input, hidden, and output layers. In this study, we examine how the information flows are shaped by the network parameters, such as depth, sparsity, weight constraints, and hidden representations. Here, we adopt autoencoders as models of deep learning, because (i) they have clear guidelines for their information flows, and (ii) they have various species, such as vanilla, sparse, tied, variational, and label autoencoders. We measured their information flows using Rényi’s matrix-based [Formula: see text]-order entropy functional. As learning progresses, they show a typical fitting phase where the amounts of input-to-hidden and hidden-to-output mutual information both increase. In the last stage of learning, however, some autoencoders show a simplifying phase, previously called the “compression phase”, where input-to-hidden mutual information diminishes. In particular, the sparsity regularization of hidden activities amplifies the simplifying phase. However, tied, variational, and label autoencoders do not have a simplifying phase. Nevertheless, all autoencoders have similar reconstruction errors for training and test data. Thus, the simplifying phase does not seem to be necessary for the generalization of learning. MDPI 2021-07-05 /pmc/articles/PMC8303402/ /pubmed/34356403 http://dx.doi.org/10.3390/e23070862 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Lee, Sungyeop
Jo, Junghyo
Information Flows of Diverse Autoencoders
title Information Flows of Diverse Autoencoders
title_full Information Flows of Diverse Autoencoders
title_fullStr Information Flows of Diverse Autoencoders
title_full_unstemmed Information Flows of Diverse Autoencoders
title_short Information Flows of Diverse Autoencoders
title_sort information flows of diverse autoencoders
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8303402/
https://www.ncbi.nlm.nih.gov/pubmed/34356403
http://dx.doi.org/10.3390/e23070862
work_keys_str_mv AT leesungyeop informationflowsofdiverseautoencoders
AT jojunghyo informationflowsofdiverseautoencoders