Cargando…

Mutual Information Gain and Linear/Nonlinear Redundancy for Agent Learning, Sequence Analysis, and Modeling

In many applications, intelligent agents need to identify any structure or apparent randomness in an environment and respond appropriately. We use the relative entropy to separate and quantify the presence of both linear and nonlinear redundancy in a sequence and we introduce the new quantities of t...

Descripción completa

Detalles Bibliográficos
Autor principal: Gibson, Jerry D.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7517148/
https://www.ncbi.nlm.nih.gov/pubmed/33286380
http://dx.doi.org/10.3390/e22060608
_version_ 1783587163144716288
author Gibson, Jerry D.
author_facet Gibson, Jerry D.
author_sort Gibson, Jerry D.
collection PubMed
description In many applications, intelligent agents need to identify any structure or apparent randomness in an environment and respond appropriately. We use the relative entropy to separate and quantify the presence of both linear and nonlinear redundancy in a sequence and we introduce the new quantities of total mutual information gain and incremental mutual information gain. We illustrate how these new quantities can be used to analyze and characterize the structures and apparent randomness for purely autoregressive sequences and for speech signals with long and short term linear redundancies. The mutual information gain is shown to be an important new tool for capturing and quantifying learning for sequence modeling and analysis.
format Online
Article
Text
id pubmed-7517148
institution National Center for Biotechnology Information
language English
publishDate 2020
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-75171482020-11-09 Mutual Information Gain and Linear/Nonlinear Redundancy for Agent Learning, Sequence Analysis, and Modeling Gibson, Jerry D. Entropy (Basel) Article In many applications, intelligent agents need to identify any structure or apparent randomness in an environment and respond appropriately. We use the relative entropy to separate and quantify the presence of both linear and nonlinear redundancy in a sequence and we introduce the new quantities of total mutual information gain and incremental mutual information gain. We illustrate how these new quantities can be used to analyze and characterize the structures and apparent randomness for purely autoregressive sequences and for speech signals with long and short term linear redundancies. The mutual information gain is shown to be an important new tool for capturing and quantifying learning for sequence modeling and analysis. MDPI 2020-05-30 /pmc/articles/PMC7517148/ /pubmed/33286380 http://dx.doi.org/10.3390/e22060608 Text en © 2020 by the author. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Gibson, Jerry D.
Mutual Information Gain and Linear/Nonlinear Redundancy for Agent Learning, Sequence Analysis, and Modeling
title Mutual Information Gain and Linear/Nonlinear Redundancy for Agent Learning, Sequence Analysis, and Modeling
title_full Mutual Information Gain and Linear/Nonlinear Redundancy for Agent Learning, Sequence Analysis, and Modeling
title_fullStr Mutual Information Gain and Linear/Nonlinear Redundancy for Agent Learning, Sequence Analysis, and Modeling
title_full_unstemmed Mutual Information Gain and Linear/Nonlinear Redundancy for Agent Learning, Sequence Analysis, and Modeling
title_short Mutual Information Gain and Linear/Nonlinear Redundancy for Agent Learning, Sequence Analysis, and Modeling
title_sort mutual information gain and linear/nonlinear redundancy for agent learning, sequence analysis, and modeling
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7517148/
https://www.ncbi.nlm.nih.gov/pubmed/33286380
http://dx.doi.org/10.3390/e22060608
work_keys_str_mv AT gibsonjerryd mutualinformationgainandlinearnonlinearredundancyforagentlearningsequenceanalysisandmodeling