Cargando…
Probabilistic Deterministic Finite Automata and Recurrent Networks, Revisited
Reservoir computers (RCs) and recurrent neural networks (RNNs) can mimic any finite-state automaton in theory, and some workers demonstrated that this can hold in practice. We test the capability of generalized linear models, RCs, and Long Short-Term Memory (LSTM) RNN architectures to predict the st...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8774624/ https://www.ncbi.nlm.nih.gov/pubmed/35052116 http://dx.doi.org/10.3390/e24010090 |
_version_ | 1784636393062924288 |
---|---|
author | Marzen, Sarah E. Crutchfield, James P. |
author_facet | Marzen, Sarah E. Crutchfield, James P. |
author_sort | Marzen, Sarah E. |
collection | PubMed |
description | Reservoir computers (RCs) and recurrent neural networks (RNNs) can mimic any finite-state automaton in theory, and some workers demonstrated that this can hold in practice. We test the capability of generalized linear models, RCs, and Long Short-Term Memory (LSTM) RNN architectures to predict the stochastic processes generated by a large suite of probabilistic deterministic finite-state automata (PDFA) in the small-data limit according to two metrics: predictive accuracy and distance to a predictive rate-distortion curve. The latter provides a sense of whether or not the RNN is a lossy predictive feature extractor in the information-theoretic sense. PDFAs provide an excellent performance benchmark in that they can be systematically enumerated, the randomness and correlation structure of their generated processes are exactly known, and their optimal memory-limited predictors are easily computed. With less data than is needed to make a good prediction, LSTMs surprisingly lose at predictive accuracy, but win at lossy predictive feature extraction. These results highlight the utility of causal states in understanding the capabilities of RNNs to predict. |
format | Online Article Text |
id | pubmed-8774624 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-87746242022-01-21 Probabilistic Deterministic Finite Automata and Recurrent Networks, Revisited Marzen, Sarah E. Crutchfield, James P. Entropy (Basel) Article Reservoir computers (RCs) and recurrent neural networks (RNNs) can mimic any finite-state automaton in theory, and some workers demonstrated that this can hold in practice. We test the capability of generalized linear models, RCs, and Long Short-Term Memory (LSTM) RNN architectures to predict the stochastic processes generated by a large suite of probabilistic deterministic finite-state automata (PDFA) in the small-data limit according to two metrics: predictive accuracy and distance to a predictive rate-distortion curve. The latter provides a sense of whether or not the RNN is a lossy predictive feature extractor in the information-theoretic sense. PDFAs provide an excellent performance benchmark in that they can be systematically enumerated, the randomness and correlation structure of their generated processes are exactly known, and their optimal memory-limited predictors are easily computed. With less data than is needed to make a good prediction, LSTMs surprisingly lose at predictive accuracy, but win at lossy predictive feature extraction. These results highlight the utility of causal states in understanding the capabilities of RNNs to predict. MDPI 2022-01-06 /pmc/articles/PMC8774624/ /pubmed/35052116 http://dx.doi.org/10.3390/e24010090 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Marzen, Sarah E. Crutchfield, James P. Probabilistic Deterministic Finite Automata and Recurrent Networks, Revisited |
title | Probabilistic Deterministic Finite Automata and Recurrent Networks, Revisited |
title_full | Probabilistic Deterministic Finite Automata and Recurrent Networks, Revisited |
title_fullStr | Probabilistic Deterministic Finite Automata and Recurrent Networks, Revisited |
title_full_unstemmed | Probabilistic Deterministic Finite Automata and Recurrent Networks, Revisited |
title_short | Probabilistic Deterministic Finite Automata and Recurrent Networks, Revisited |
title_sort | probabilistic deterministic finite automata and recurrent networks, revisited |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8774624/ https://www.ncbi.nlm.nih.gov/pubmed/35052116 http://dx.doi.org/10.3390/e24010090 |
work_keys_str_mv | AT marzensarahe probabilisticdeterministicfiniteautomataandrecurrentnetworksrevisited AT crutchfieldjamesp probabilisticdeterministicfiniteautomataandrecurrentnetworksrevisited |