Cargando…

Digital computing through randomness and order in neural networks

We propose that coding and decoding in the brain are achieved through digital computation using three principles: relative ordinal coding of inputs, random connections between neurons, and belief voting. Due to randomization and despite the coarseness of the relative codes, we show that these princi...

Descripción completa

Detalles Bibliográficos
Autores principales: Pitti, Alexandre, Weidmann, Claudio, Quoy, Mathias
Formato: Online Artículo Texto
Lenguaje:English
Publicado: National Academy of Sciences 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9388095/
https://www.ncbi.nlm.nih.gov/pubmed/35947616
http://dx.doi.org/10.1073/pnas.2115335119
_version_ 1784770147285729280
author Pitti, Alexandre
Weidmann, Claudio
Quoy, Mathias
author_facet Pitti, Alexandre
Weidmann, Claudio
Quoy, Mathias
author_sort Pitti, Alexandre
collection PubMed
description We propose that coding and decoding in the brain are achieved through digital computation using three principles: relative ordinal coding of inputs, random connections between neurons, and belief voting. Due to randomization and despite the coarseness of the relative codes, we show that these principles are sufficient for coding and decoding sequences with error-free reconstruction. In particular, the number of neurons needed grows linearly with the size of the input repertoire growing exponentially. We illustrate our model by reconstructing sequences with repertoires on the order of a billion items. From this, we derive the Shannon equations for the capacity limit to learn and transfer information in the neural population, which is then generalized to any type of neural network. Following the maximum entropy principle of efficient coding, we show that random connections serve to decorrelate redundant information in incoming signals, creating more compact codes for neurons and therefore, conveying a larger amount of information. Henceforth, despite the unreliability of the relative codes, few neurons become necessary to discriminate the original signal without error. Finally, we discuss the significance of this digital computation model regarding neurobiological findings in the brain and more generally with artificial intelligence algorithms, with a view toward a neural information theory and the design of digital neural networks.
format Online
Article
Text
id pubmed-9388095
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher National Academy of Sciences
record_format MEDLINE/PubMed
spelling pubmed-93880952023-02-10 Digital computing through randomness and order in neural networks Pitti, Alexandre Weidmann, Claudio Quoy, Mathias Proc Natl Acad Sci U S A Biological Sciences We propose that coding and decoding in the brain are achieved through digital computation using three principles: relative ordinal coding of inputs, random connections between neurons, and belief voting. Due to randomization and despite the coarseness of the relative codes, we show that these principles are sufficient for coding and decoding sequences with error-free reconstruction. In particular, the number of neurons needed grows linearly with the size of the input repertoire growing exponentially. We illustrate our model by reconstructing sequences with repertoires on the order of a billion items. From this, we derive the Shannon equations for the capacity limit to learn and transfer information in the neural population, which is then generalized to any type of neural network. Following the maximum entropy principle of efficient coding, we show that random connections serve to decorrelate redundant information in incoming signals, creating more compact codes for neurons and therefore, conveying a larger amount of information. Henceforth, despite the unreliability of the relative codes, few neurons become necessary to discriminate the original signal without error. Finally, we discuss the significance of this digital computation model regarding neurobiological findings in the brain and more generally with artificial intelligence algorithms, with a view toward a neural information theory and the design of digital neural networks. National Academy of Sciences 2022-08-10 2022-08-16 /pmc/articles/PMC9388095/ /pubmed/35947616 http://dx.doi.org/10.1073/pnas.2115335119 Text en Copyright © 2022 the Author(s). Published by PNAS. https://creativecommons.org/licenses/by-nc-nd/4.0/This article is distributed under Creative Commons Attribution-NonCommercial-NoDerivatives License 4.0 (CC BY-NC-ND) (https://creativecommons.org/licenses/by-nc-nd/4.0/) .
spellingShingle Biological Sciences
Pitti, Alexandre
Weidmann, Claudio
Quoy, Mathias
Digital computing through randomness and order in neural networks
title Digital computing through randomness and order in neural networks
title_full Digital computing through randomness and order in neural networks
title_fullStr Digital computing through randomness and order in neural networks
title_full_unstemmed Digital computing through randomness and order in neural networks
title_short Digital computing through randomness and order in neural networks
title_sort digital computing through randomness and order in neural networks
topic Biological Sciences
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9388095/
https://www.ncbi.nlm.nih.gov/pubmed/35947616
http://dx.doi.org/10.1073/pnas.2115335119
work_keys_str_mv AT pittialexandre digitalcomputingthroughrandomnessandorderinneuralnetworks
AT weidmannclaudio digitalcomputingthroughrandomnessandorderinneuralnetworks
AT quoymathias digitalcomputingthroughrandomnessandorderinneuralnetworks