Cargando…

Learning probabilistic neural representations with randomly connected circuits

The brain represents and reasons probabilistically about complex stimuli and motor actions using a noisy, spike-based neural code. A key building block for such neural computations, as well as the basis for supervised and unsupervised learning, is the ability to estimate the surprise or likelihood o...

Descripción completa

Detalles Bibliográficos
Autores principales: Maoz, Ori, Tkačik, Gašper, Esteki, Mohamad Saleh, Kiani, Roozbeh, Schneidman, Elad
Formato: Online Artículo Texto
Lenguaje:English
Publicado: National Academy of Sciences 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7547210/
https://www.ncbi.nlm.nih.gov/pubmed/32948691
http://dx.doi.org/10.1073/pnas.1912804117
_version_ 1783592379924611072
author Maoz, Ori
Tkačik, Gašper
Esteki, Mohamad Saleh
Kiani, Roozbeh
Schneidman, Elad
author_facet Maoz, Ori
Tkačik, Gašper
Esteki, Mohamad Saleh
Kiani, Roozbeh
Schneidman, Elad
author_sort Maoz, Ori
collection PubMed
description The brain represents and reasons probabilistically about complex stimuli and motor actions using a noisy, spike-based neural code. A key building block for such neural computations, as well as the basis for supervised and unsupervised learning, is the ability to estimate the surprise or likelihood of incoming high-dimensional neural activity patterns. Despite progress in statistical modeling of neural responses and deep learning, current approaches either do not scale to large neural populations or cannot be implemented using biologically realistic mechanisms. Inspired by the sparse and random connectivity of real neuronal circuits, we present a model for neural codes that accurately estimates the likelihood of individual spiking patterns and has a straightforward, scalable, efficient, learnable, and realistic neural implementation. This model’s performance on simultaneously recorded spiking activity of >100 neurons in the monkey visual and prefrontal cortices is comparable with or better than that of state-of-the-art models. Importantly, the model can be learned using a small number of samples and using a local learning rule that utilizes noise intrinsic to neural circuits. Slower, structural changes in random connectivity, consistent with rewiring and pruning processes, further improve the efficiency and sparseness of the resulting neural representations. Our results merge insights from neuroanatomy, machine learning, and theoretical neuroscience to suggest random sparse connectivity as a key design principle for neuronal computation.
format Online
Article
Text
id pubmed-7547210
institution National Center for Biotechnology Information
language English
publishDate 2020
publisher National Academy of Sciences
record_format MEDLINE/PubMed
spelling pubmed-75472102020-10-22 Learning probabilistic neural representations with randomly connected circuits Maoz, Ori Tkačik, Gašper Esteki, Mohamad Saleh Kiani, Roozbeh Schneidman, Elad Proc Natl Acad Sci U S A Biological Sciences The brain represents and reasons probabilistically about complex stimuli and motor actions using a noisy, spike-based neural code. A key building block for such neural computations, as well as the basis for supervised and unsupervised learning, is the ability to estimate the surprise or likelihood of incoming high-dimensional neural activity patterns. Despite progress in statistical modeling of neural responses and deep learning, current approaches either do not scale to large neural populations or cannot be implemented using biologically realistic mechanisms. Inspired by the sparse and random connectivity of real neuronal circuits, we present a model for neural codes that accurately estimates the likelihood of individual spiking patterns and has a straightforward, scalable, efficient, learnable, and realistic neural implementation. This model’s performance on simultaneously recorded spiking activity of >100 neurons in the monkey visual and prefrontal cortices is comparable with or better than that of state-of-the-art models. Importantly, the model can be learned using a small number of samples and using a local learning rule that utilizes noise intrinsic to neural circuits. Slower, structural changes in random connectivity, consistent with rewiring and pruning processes, further improve the efficiency and sparseness of the resulting neural representations. Our results merge insights from neuroanatomy, machine learning, and theoretical neuroscience to suggest random sparse connectivity as a key design principle for neuronal computation. National Academy of Sciences 2020-10-06 2020-09-18 /pmc/articles/PMC7547210/ /pubmed/32948691 http://dx.doi.org/10.1073/pnas.1912804117 Text en Copyright © 2020 the Author(s). Published by PNAS. https://creativecommons.org/licenses/by-nc-nd/4.0/ https://creativecommons.org/licenses/by-nc-nd/4.0/This open access article is distributed under Creative Commons Attribution-NonCommercial-NoDerivatives License 4.0 (CC BY-NC-ND) (https://creativecommons.org/licenses/by-nc-nd/4.0/) .
spellingShingle Biological Sciences
Maoz, Ori
Tkačik, Gašper
Esteki, Mohamad Saleh
Kiani, Roozbeh
Schneidman, Elad
Learning probabilistic neural representations with randomly connected circuits
title Learning probabilistic neural representations with randomly connected circuits
title_full Learning probabilistic neural representations with randomly connected circuits
title_fullStr Learning probabilistic neural representations with randomly connected circuits
title_full_unstemmed Learning probabilistic neural representations with randomly connected circuits
title_short Learning probabilistic neural representations with randomly connected circuits
title_sort learning probabilistic neural representations with randomly connected circuits
topic Biological Sciences
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7547210/
https://www.ncbi.nlm.nih.gov/pubmed/32948691
http://dx.doi.org/10.1073/pnas.1912804117
work_keys_str_mv AT maozori learningprobabilisticneuralrepresentationswithrandomlyconnectedcircuits
AT tkacikgasper learningprobabilisticneuralrepresentationswithrandomlyconnectedcircuits
AT estekimohamadsaleh learningprobabilisticneuralrepresentationswithrandomlyconnectedcircuits
AT kianiroozbeh learningprobabilisticneuralrepresentationswithrandomlyconnectedcircuits
AT schneidmanelad learningprobabilisticneuralrepresentationswithrandomlyconnectedcircuits