Cargando…
Partial Autoinformation to Characterize Symbolic Sequences
An information-theoretic approach to numerically determine the Markov order of discrete stochastic processes defined over a finite state space is introduced. To measure statistical dependencies between different time points of symbolic time series, two information-theoretic measures are proposed. Th...
Autor principal: | |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2018
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6194330/ https://www.ncbi.nlm.nih.gov/pubmed/30369884 http://dx.doi.org/10.3389/fphys.2018.01382 |
_version_ | 1783364216923619328 |
---|---|
author | von Wegner, Frederic |
author_facet | von Wegner, Frederic |
author_sort | von Wegner, Frederic |
collection | PubMed |
description | An information-theoretic approach to numerically determine the Markov order of discrete stochastic processes defined over a finite state space is introduced. To measure statistical dependencies between different time points of symbolic time series, two information-theoretic measures are proposed. The first measure is time-lagged mutual information between the random variables X(n) and X(n+k), representing the values of the process at time points n and n + k, respectively. The measure will be termed autoinformation, in analogy to the autocorrelation function for metric time series, but using Shannon entropy rather than linear correlation. This measure is complemented by the conditional mutual information between X(n) and X(n+k), removing the influence of the intermediate values X(n+k−1), …, X(n+1). The second measure is termed partial autoinformation, in analogy to the partial autocorrelation function (PACF) in metric time series analysis. Mathematical relations with known quantities such as the entropy rate and active information storage are established. Both measures are applied to a number of examples, ranging from theoretical Markov and non-Markov processes with known stochastic properties, to models from statistical physics, and finally, to a discrete transform of an EEG data set. The combination of autoinformation and partial autoinformation yields important insights into the temporal structure of the data in all test cases. For first- and higher-order Markov processes, partial autoinformation correctly identifies the order parameter, but also suggests extended, non-Markovian effects in the examples that lack the Markov property. For three hidden Markov models (HMMs), the underlying Markov order is found. The combination of both quantities may be used as an early step in the analysis of experimental, non-metric time series and can be employed to discover higher-order Markov dependencies, non-Markovianity and periodicities in symbolic time series. |
format | Online Article Text |
id | pubmed-6194330 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2018 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-61943302018-10-26 Partial Autoinformation to Characterize Symbolic Sequences von Wegner, Frederic Front Physiol Physiology An information-theoretic approach to numerically determine the Markov order of discrete stochastic processes defined over a finite state space is introduced. To measure statistical dependencies between different time points of symbolic time series, two information-theoretic measures are proposed. The first measure is time-lagged mutual information between the random variables X(n) and X(n+k), representing the values of the process at time points n and n + k, respectively. The measure will be termed autoinformation, in analogy to the autocorrelation function for metric time series, but using Shannon entropy rather than linear correlation. This measure is complemented by the conditional mutual information between X(n) and X(n+k), removing the influence of the intermediate values X(n+k−1), …, X(n+1). The second measure is termed partial autoinformation, in analogy to the partial autocorrelation function (PACF) in metric time series analysis. Mathematical relations with known quantities such as the entropy rate and active information storage are established. Both measures are applied to a number of examples, ranging from theoretical Markov and non-Markov processes with known stochastic properties, to models from statistical physics, and finally, to a discrete transform of an EEG data set. The combination of autoinformation and partial autoinformation yields important insights into the temporal structure of the data in all test cases. For first- and higher-order Markov processes, partial autoinformation correctly identifies the order parameter, but also suggests extended, non-Markovian effects in the examples that lack the Markov property. For three hidden Markov models (HMMs), the underlying Markov order is found. The combination of both quantities may be used as an early step in the analysis of experimental, non-metric time series and can be employed to discover higher-order Markov dependencies, non-Markovianity and periodicities in symbolic time series. Frontiers Media S.A. 2018-10-11 /pmc/articles/PMC6194330/ /pubmed/30369884 http://dx.doi.org/10.3389/fphys.2018.01382 Text en Copyright © 2018 von Wegner. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Physiology von Wegner, Frederic Partial Autoinformation to Characterize Symbolic Sequences |
title | Partial Autoinformation to Characterize Symbolic Sequences |
title_full | Partial Autoinformation to Characterize Symbolic Sequences |
title_fullStr | Partial Autoinformation to Characterize Symbolic Sequences |
title_full_unstemmed | Partial Autoinformation to Characterize Symbolic Sequences |
title_short | Partial Autoinformation to Characterize Symbolic Sequences |
title_sort | partial autoinformation to characterize symbolic sequences |
topic | Physiology |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6194330/ https://www.ncbi.nlm.nih.gov/pubmed/30369884 http://dx.doi.org/10.3389/fphys.2018.01382 |
work_keys_str_mv | AT vonwegnerfrederic partialautoinformationtocharacterizesymbolicsequences |