Cargando…

Dissecting Deep Learning Networks—Visualizing Mutual Information

Deep Learning (DL) networks are recent revolutionary developments in artificial intelligence research. Typical networks are stacked by groups of layers that are further composed of many convolutional kernels or neurons. In network design, many hyper-parameters need to be defined heuristically before...

Descripción completa

Detalles Bibliográficos
Autores principales: Fang, Hui, Wang, Victoria, Yamaguchi, Motonori
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2018
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7512386/
https://www.ncbi.nlm.nih.gov/pubmed/33266547
http://dx.doi.org/10.3390/e20110823
_version_ 1783586145887584256
author Fang, Hui
Wang, Victoria
Yamaguchi, Motonori
author_facet Fang, Hui
Wang, Victoria
Yamaguchi, Motonori
author_sort Fang, Hui
collection PubMed
description Deep Learning (DL) networks are recent revolutionary developments in artificial intelligence research. Typical networks are stacked by groups of layers that are further composed of many convolutional kernels or neurons. In network design, many hyper-parameters need to be defined heuristically before training in order to achieve high cross-validation accuracies. However, accuracy evaluation from the output layer alone is not sufficient to specify the roles of the hidden units in associated networks. This results in a significant knowledge gap between DL’s wider applications and its limited theoretical understanding. To narrow the knowledge gap, our study explores visualization techniques to illustrate the mutual information (MI) in DL networks. The MI is a theoretical measurement, reflecting the relationship between two sets of random variables even if their relationship is highly non-linear and hidden in high-dimensional data. Our study aims to understand the roles of DL units in classification performance of the networks. Via a series of experiments using several popular DL networks, it shows that the visualization of MI and its change patterns between the input/output with the hidden layers and basic units can facilitate a better understanding of these DL units’ roles. Our investigation on network convergence suggests a more objective manner to potentially evaluate DL networks. Furthermore, the visualization provides a useful tool to gain insights into the network performance, and thus to potentially facilitate the design of better network architectures by identifying redundancy and less-effective network units.
format Online
Article
Text
id pubmed-7512386
institution National Center for Biotechnology Information
language English
publishDate 2018
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-75123862020-11-09 Dissecting Deep Learning Networks—Visualizing Mutual Information Fang, Hui Wang, Victoria Yamaguchi, Motonori Entropy (Basel) Article Deep Learning (DL) networks are recent revolutionary developments in artificial intelligence research. Typical networks are stacked by groups of layers that are further composed of many convolutional kernels or neurons. In network design, many hyper-parameters need to be defined heuristically before training in order to achieve high cross-validation accuracies. However, accuracy evaluation from the output layer alone is not sufficient to specify the roles of the hidden units in associated networks. This results in a significant knowledge gap between DL’s wider applications and its limited theoretical understanding. To narrow the knowledge gap, our study explores visualization techniques to illustrate the mutual information (MI) in DL networks. The MI is a theoretical measurement, reflecting the relationship between two sets of random variables even if their relationship is highly non-linear and hidden in high-dimensional data. Our study aims to understand the roles of DL units in classification performance of the networks. Via a series of experiments using several popular DL networks, it shows that the visualization of MI and its change patterns between the input/output with the hidden layers and basic units can facilitate a better understanding of these DL units’ roles. Our investigation on network convergence suggests a more objective manner to potentially evaluate DL networks. Furthermore, the visualization provides a useful tool to gain insights into the network performance, and thus to potentially facilitate the design of better network architectures by identifying redundancy and less-effective network units. MDPI 2018-10-26 /pmc/articles/PMC7512386/ /pubmed/33266547 http://dx.doi.org/10.3390/e20110823 Text en © 2018 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Fang, Hui
Wang, Victoria
Yamaguchi, Motonori
Dissecting Deep Learning Networks—Visualizing Mutual Information
title Dissecting Deep Learning Networks—Visualizing Mutual Information
title_full Dissecting Deep Learning Networks—Visualizing Mutual Information
title_fullStr Dissecting Deep Learning Networks—Visualizing Mutual Information
title_full_unstemmed Dissecting Deep Learning Networks—Visualizing Mutual Information
title_short Dissecting Deep Learning Networks—Visualizing Mutual Information
title_sort dissecting deep learning networks—visualizing mutual information
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7512386/
https://www.ncbi.nlm.nih.gov/pubmed/33266547
http://dx.doi.org/10.3390/e20110823
work_keys_str_mv AT fanghui dissectingdeeplearningnetworksvisualizingmutualinformation
AT wangvictoria dissectingdeeplearningnetworksvisualizingmutualinformation
AT yamaguchimotonori dissectingdeeplearningnetworksvisualizingmutualinformation