Cargando…

A Review of Shannon and Differential Entropy Rate Estimation

In this paper, we present a review of Shannon and differential entropy rate estimation techniques. Entropy rate, which measures the average information gain from a stochastic process, is a measure of uncertainty and complexity of a stochastic process. We discuss the estimation of entropy rate from e...

Descripción completa

Detalles Bibliográficos
Autores principales: Feutrill, Andrew, Roughan, Matthew
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8392187/
https://www.ncbi.nlm.nih.gov/pubmed/34441186
http://dx.doi.org/10.3390/e23081046
_version_ 1783743442259542016
author Feutrill, Andrew
Roughan, Matthew
author_facet Feutrill, Andrew
Roughan, Matthew
author_sort Feutrill, Andrew
collection PubMed
description In this paper, we present a review of Shannon and differential entropy rate estimation techniques. Entropy rate, which measures the average information gain from a stochastic process, is a measure of uncertainty and complexity of a stochastic process. We discuss the estimation of entropy rate from empirical data, and review both parametric and non-parametric techniques. We look at many different assumptions on properties of the processes for parametric processes, in particular focussing on Markov and Gaussian assumptions. Non-parametric estimation relies on limit theorems which involve the entropy rate from observations, and to discuss these, we introduce some theory and the practical implementations of estimators of this type.
format Online
Article
Text
id pubmed-8392187
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-83921872021-08-28 A Review of Shannon and Differential Entropy Rate Estimation Feutrill, Andrew Roughan, Matthew Entropy (Basel) Review In this paper, we present a review of Shannon and differential entropy rate estimation techniques. Entropy rate, which measures the average information gain from a stochastic process, is a measure of uncertainty and complexity of a stochastic process. We discuss the estimation of entropy rate from empirical data, and review both parametric and non-parametric techniques. We look at many different assumptions on properties of the processes for parametric processes, in particular focussing on Markov and Gaussian assumptions. Non-parametric estimation relies on limit theorems which involve the entropy rate from observations, and to discuss these, we introduce some theory and the practical implementations of estimators of this type. MDPI 2021-08-13 /pmc/articles/PMC8392187/ /pubmed/34441186 http://dx.doi.org/10.3390/e23081046 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Review
Feutrill, Andrew
Roughan, Matthew
A Review of Shannon and Differential Entropy Rate Estimation
title A Review of Shannon and Differential Entropy Rate Estimation
title_full A Review of Shannon and Differential Entropy Rate Estimation
title_fullStr A Review of Shannon and Differential Entropy Rate Estimation
title_full_unstemmed A Review of Shannon and Differential Entropy Rate Estimation
title_short A Review of Shannon and Differential Entropy Rate Estimation
title_sort review of shannon and differential entropy rate estimation
topic Review
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8392187/
https://www.ncbi.nlm.nih.gov/pubmed/34441186
http://dx.doi.org/10.3390/e23081046
work_keys_str_mv AT feutrillandrew areviewofshannonanddifferentialentropyrateestimation
AT roughanmatthew areviewofshannonanddifferentialentropyrateestimation
AT feutrillandrew reviewofshannonanddifferentialentropyrateestimation
AT roughanmatthew reviewofshannonanddifferentialentropyrateestimation