Cargando…

Learning in Feedforward Neural Networks Accelerated by Transfer Entropy

Current neural networks architectures are many times harder to train because of the increasing size and complexity of the used datasets. Our objective is to design more efficient training algorithms utilizing causal relationships inferred from neural networks. The transfer entropy (TE) was initially...

Descripción completa

Detalles Bibliográficos
Autores principales: Moldovan, Adrian, Caţaron, Angel, Andonie, Răzvan
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7516405/
https://www.ncbi.nlm.nih.gov/pubmed/33285877
http://dx.doi.org/10.3390/e22010102
_version_ 1783586993470439424
author Moldovan, Adrian
Caţaron, Angel
Andonie, Răzvan
author_facet Moldovan, Adrian
Caţaron, Angel
Andonie, Răzvan
author_sort Moldovan, Adrian
collection PubMed
description Current neural networks architectures are many times harder to train because of the increasing size and complexity of the used datasets. Our objective is to design more efficient training algorithms utilizing causal relationships inferred from neural networks. The transfer entropy (TE) was initially introduced as an information transfer measure used to quantify the statistical coherence between events (time series). Later, it was related to causality, even if they are not the same. There are only few papers reporting applications of causality or TE in neural networks. Our contribution is an information-theoretical method for analyzing information transfer between the nodes of feedforward neural networks. The information transfer is measured by the TE of feedback neural connections. Intuitively, TE measures the relevance of a connection in the network and the feedback amplifies this connection. We introduce a backpropagation type training algorithm that uses TE feedback connections to improve its performance.
format Online
Article
Text
id pubmed-7516405
institution National Center for Biotechnology Information
language English
publishDate 2020
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-75164052020-11-09 Learning in Feedforward Neural Networks Accelerated by Transfer Entropy Moldovan, Adrian Caţaron, Angel Andonie, Răzvan Entropy (Basel) Article Current neural networks architectures are many times harder to train because of the increasing size and complexity of the used datasets. Our objective is to design more efficient training algorithms utilizing causal relationships inferred from neural networks. The transfer entropy (TE) was initially introduced as an information transfer measure used to quantify the statistical coherence between events (time series). Later, it was related to causality, even if they are not the same. There are only few papers reporting applications of causality or TE in neural networks. Our contribution is an information-theoretical method for analyzing information transfer between the nodes of feedforward neural networks. The information transfer is measured by the TE of feedback neural connections. Intuitively, TE measures the relevance of a connection in the network and the feedback amplifies this connection. We introduce a backpropagation type training algorithm that uses TE feedback connections to improve its performance. MDPI 2020-01-16 /pmc/articles/PMC7516405/ /pubmed/33285877 http://dx.doi.org/10.3390/e22010102 Text en © 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Moldovan, Adrian
Caţaron, Angel
Andonie, Răzvan
Learning in Feedforward Neural Networks Accelerated by Transfer Entropy
title Learning in Feedforward Neural Networks Accelerated by Transfer Entropy
title_full Learning in Feedforward Neural Networks Accelerated by Transfer Entropy
title_fullStr Learning in Feedforward Neural Networks Accelerated by Transfer Entropy
title_full_unstemmed Learning in Feedforward Neural Networks Accelerated by Transfer Entropy
title_short Learning in Feedforward Neural Networks Accelerated by Transfer Entropy
title_sort learning in feedforward neural networks accelerated by transfer entropy
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7516405/
https://www.ncbi.nlm.nih.gov/pubmed/33285877
http://dx.doi.org/10.3390/e22010102
work_keys_str_mv AT moldovanadrian learninginfeedforwardneuralnetworksacceleratedbytransferentropy
AT cataronangel learninginfeedforwardneuralnetworksacceleratedbytransferentropy
AT andonierazvan learninginfeedforwardneuralnetworksacceleratedbytransferentropy