Cargando…

Identification of probabilities

Within psychology, neuroscience and artificial intelligence, there has been increasing interest in the proposal that the brain builds probabilistic models of sensory and linguistic input: that is, to infer a probabilistic model from a sample. The practical problems of such inference are substantial:...

Descripción completa

Detalles Bibliográficos
Autores principales: Vitányi, Paul M.B., Chater, Nick
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Academic Press 2017
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5341760/
https://www.ncbi.nlm.nih.gov/pubmed/28298701
http://dx.doi.org/10.1016/j.jmp.2016.11.004
_version_ 1782513033682092032
author Vitányi, Paul M.B.
Chater, Nick
author_facet Vitányi, Paul M.B.
Chater, Nick
author_sort Vitányi, Paul M.B.
collection PubMed
description Within psychology, neuroscience and artificial intelligence, there has been increasing interest in the proposal that the brain builds probabilistic models of sensory and linguistic input: that is, to infer a probabilistic model from a sample. The practical problems of such inference are substantial: the brain has limited data and restricted computational resources. But there is a more fundamental question: is the problem of inferring a probabilistic model from a sample possible even in principle? We explore this question and find some surprisingly positive and general results. First, for a broad class of probability distributions characterized by computability restrictions, we specify a learning algorithm that will almost surely identify a probability distribution in the limit given a finite i.i.d. sample of sufficient but unknown length. This is similarly shown to hold for sequences generated by a broad class of Markov chains, subject to computability assumptions. The technical tool is the strong law of large numbers. Second, for a large class of dependent sequences, we specify an algorithm which identifies in the limit a computable measure for which the sequence is typical, in the sense of Martin-Löf (there may be more than one such measure). The technical tool is the theory of Kolmogorov complexity. We analyze the associated predictions in both cases. We also briefly consider special cases, including language learning, and wider theoretical implications for psychology.
format Online
Article
Text
id pubmed-5341760
institution National Center for Biotechnology Information
language English
publishDate 2017
publisher Academic Press
record_format MEDLINE/PubMed
spelling pubmed-53417602017-03-13 Identification of probabilities Vitányi, Paul M.B. Chater, Nick J Math Psychol Article Within psychology, neuroscience and artificial intelligence, there has been increasing interest in the proposal that the brain builds probabilistic models of sensory and linguistic input: that is, to infer a probabilistic model from a sample. The practical problems of such inference are substantial: the brain has limited data and restricted computational resources. But there is a more fundamental question: is the problem of inferring a probabilistic model from a sample possible even in principle? We explore this question and find some surprisingly positive and general results. First, for a broad class of probability distributions characterized by computability restrictions, we specify a learning algorithm that will almost surely identify a probability distribution in the limit given a finite i.i.d. sample of sufficient but unknown length. This is similarly shown to hold for sequences generated by a broad class of Markov chains, subject to computability assumptions. The technical tool is the strong law of large numbers. Second, for a large class of dependent sequences, we specify an algorithm which identifies in the limit a computable measure for which the sequence is typical, in the sense of Martin-Löf (there may be more than one such measure). The technical tool is the theory of Kolmogorov complexity. We analyze the associated predictions in both cases. We also briefly consider special cases, including language learning, and wider theoretical implications for psychology. Academic Press 2017-02 /pmc/articles/PMC5341760/ /pubmed/28298701 http://dx.doi.org/10.1016/j.jmp.2016.11.004 Text en © 2016 The Author(s) http://creativecommons.org/licenses/by-nc-nd/4.0/ This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/).
spellingShingle Article
Vitányi, Paul M.B.
Chater, Nick
Identification of probabilities
title Identification of probabilities
title_full Identification of probabilities
title_fullStr Identification of probabilities
title_full_unstemmed Identification of probabilities
title_short Identification of probabilities
title_sort identification of probabilities
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5341760/
https://www.ncbi.nlm.nih.gov/pubmed/28298701
http://dx.doi.org/10.1016/j.jmp.2016.11.004
work_keys_str_mv AT vitanyipaulmb identificationofprobabilities
AT chaternick identificationofprobabilities