Cargando…

Probabilistic Deterministic Finite Automata and Recurrent Networks, Revisited

Reservoir computers (RCs) and recurrent neural networks (RNNs) can mimic any finite-state automaton in theory, and some workers demonstrated that this can hold in practice. We test the capability of generalized linear models, RCs, and Long Short-Term Memory (LSTM) RNN architectures to predict the st...

Descripción completa

Detalles Bibliográficos
Autores principales: Marzen, Sarah E., Crutchfield, James P.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8774624/
https://www.ncbi.nlm.nih.gov/pubmed/35052116
http://dx.doi.org/10.3390/e24010090
_version_ 1784636393062924288
author Marzen, Sarah E.
Crutchfield, James P.
author_facet Marzen, Sarah E.
Crutchfield, James P.
author_sort Marzen, Sarah E.
collection PubMed
description Reservoir computers (RCs) and recurrent neural networks (RNNs) can mimic any finite-state automaton in theory, and some workers demonstrated that this can hold in practice. We test the capability of generalized linear models, RCs, and Long Short-Term Memory (LSTM) RNN architectures to predict the stochastic processes generated by a large suite of probabilistic deterministic finite-state automata (PDFA) in the small-data limit according to two metrics: predictive accuracy and distance to a predictive rate-distortion curve. The latter provides a sense of whether or not the RNN is a lossy predictive feature extractor in the information-theoretic sense. PDFAs provide an excellent performance benchmark in that they can be systematically enumerated, the randomness and correlation structure of their generated processes are exactly known, and their optimal memory-limited predictors are easily computed. With less data than is needed to make a good prediction, LSTMs surprisingly lose at predictive accuracy, but win at lossy predictive feature extraction. These results highlight the utility of causal states in understanding the capabilities of RNNs to predict.
format Online
Article
Text
id pubmed-8774624
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-87746242022-01-21 Probabilistic Deterministic Finite Automata and Recurrent Networks, Revisited Marzen, Sarah E. Crutchfield, James P. Entropy (Basel) Article Reservoir computers (RCs) and recurrent neural networks (RNNs) can mimic any finite-state automaton in theory, and some workers demonstrated that this can hold in practice. We test the capability of generalized linear models, RCs, and Long Short-Term Memory (LSTM) RNN architectures to predict the stochastic processes generated by a large suite of probabilistic deterministic finite-state automata (PDFA) in the small-data limit according to two metrics: predictive accuracy and distance to a predictive rate-distortion curve. The latter provides a sense of whether or not the RNN is a lossy predictive feature extractor in the information-theoretic sense. PDFAs provide an excellent performance benchmark in that they can be systematically enumerated, the randomness and correlation structure of their generated processes are exactly known, and their optimal memory-limited predictors are easily computed. With less data than is needed to make a good prediction, LSTMs surprisingly lose at predictive accuracy, but win at lossy predictive feature extraction. These results highlight the utility of causal states in understanding the capabilities of RNNs to predict. MDPI 2022-01-06 /pmc/articles/PMC8774624/ /pubmed/35052116 http://dx.doi.org/10.3390/e24010090 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Marzen, Sarah E.
Crutchfield, James P.
Probabilistic Deterministic Finite Automata and Recurrent Networks, Revisited
title Probabilistic Deterministic Finite Automata and Recurrent Networks, Revisited
title_full Probabilistic Deterministic Finite Automata and Recurrent Networks, Revisited
title_fullStr Probabilistic Deterministic Finite Automata and Recurrent Networks, Revisited
title_full_unstemmed Probabilistic Deterministic Finite Automata and Recurrent Networks, Revisited
title_short Probabilistic Deterministic Finite Automata and Recurrent Networks, Revisited
title_sort probabilistic deterministic finite automata and recurrent networks, revisited
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8774624/
https://www.ncbi.nlm.nih.gov/pubmed/35052116
http://dx.doi.org/10.3390/e24010090
work_keys_str_mv AT marzensarahe probabilisticdeterministicfiniteautomataandrecurrentnetworksrevisited
AT crutchfieldjamesp probabilisticdeterministicfiniteautomataandrecurrentnetworksrevisited