Cargando…
Digital computing through randomness and order in neural networks
We propose that coding and decoding in the brain are achieved through digital computation using three principles: relative ordinal coding of inputs, random connections between neurons, and belief voting. Due to randomization and despite the coarseness of the relative codes, we show that these princi...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
National Academy of Sciences
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9388095/ https://www.ncbi.nlm.nih.gov/pubmed/35947616 http://dx.doi.org/10.1073/pnas.2115335119 |
_version_ | 1784770147285729280 |
---|---|
author | Pitti, Alexandre Weidmann, Claudio Quoy, Mathias |
author_facet | Pitti, Alexandre Weidmann, Claudio Quoy, Mathias |
author_sort | Pitti, Alexandre |
collection | PubMed |
description | We propose that coding and decoding in the brain are achieved through digital computation using three principles: relative ordinal coding of inputs, random connections between neurons, and belief voting. Due to randomization and despite the coarseness of the relative codes, we show that these principles are sufficient for coding and decoding sequences with error-free reconstruction. In particular, the number of neurons needed grows linearly with the size of the input repertoire growing exponentially. We illustrate our model by reconstructing sequences with repertoires on the order of a billion items. From this, we derive the Shannon equations for the capacity limit to learn and transfer information in the neural population, which is then generalized to any type of neural network. Following the maximum entropy principle of efficient coding, we show that random connections serve to decorrelate redundant information in incoming signals, creating more compact codes for neurons and therefore, conveying a larger amount of information. Henceforth, despite the unreliability of the relative codes, few neurons become necessary to discriminate the original signal without error. Finally, we discuss the significance of this digital computation model regarding neurobiological findings in the brain and more generally with artificial intelligence algorithms, with a view toward a neural information theory and the design of digital neural networks. |
format | Online Article Text |
id | pubmed-9388095 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | National Academy of Sciences |
record_format | MEDLINE/PubMed |
spelling | pubmed-93880952023-02-10 Digital computing through randomness and order in neural networks Pitti, Alexandre Weidmann, Claudio Quoy, Mathias Proc Natl Acad Sci U S A Biological Sciences We propose that coding and decoding in the brain are achieved through digital computation using three principles: relative ordinal coding of inputs, random connections between neurons, and belief voting. Due to randomization and despite the coarseness of the relative codes, we show that these principles are sufficient for coding and decoding sequences with error-free reconstruction. In particular, the number of neurons needed grows linearly with the size of the input repertoire growing exponentially. We illustrate our model by reconstructing sequences with repertoires on the order of a billion items. From this, we derive the Shannon equations for the capacity limit to learn and transfer information in the neural population, which is then generalized to any type of neural network. Following the maximum entropy principle of efficient coding, we show that random connections serve to decorrelate redundant information in incoming signals, creating more compact codes for neurons and therefore, conveying a larger amount of information. Henceforth, despite the unreliability of the relative codes, few neurons become necessary to discriminate the original signal without error. Finally, we discuss the significance of this digital computation model regarding neurobiological findings in the brain and more generally with artificial intelligence algorithms, with a view toward a neural information theory and the design of digital neural networks. National Academy of Sciences 2022-08-10 2022-08-16 /pmc/articles/PMC9388095/ /pubmed/35947616 http://dx.doi.org/10.1073/pnas.2115335119 Text en Copyright © 2022 the Author(s). Published by PNAS. https://creativecommons.org/licenses/by-nc-nd/4.0/This article is distributed under Creative Commons Attribution-NonCommercial-NoDerivatives License 4.0 (CC BY-NC-ND) (https://creativecommons.org/licenses/by-nc-nd/4.0/) . |
spellingShingle | Biological Sciences Pitti, Alexandre Weidmann, Claudio Quoy, Mathias Digital computing through randomness and order in neural networks |
title | Digital computing through randomness and order in neural networks |
title_full | Digital computing through randomness and order in neural networks |
title_fullStr | Digital computing through randomness and order in neural networks |
title_full_unstemmed | Digital computing through randomness and order in neural networks |
title_short | Digital computing through randomness and order in neural networks |
title_sort | digital computing through randomness and order in neural networks |
topic | Biological Sciences |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9388095/ https://www.ncbi.nlm.nih.gov/pubmed/35947616 http://dx.doi.org/10.1073/pnas.2115335119 |
work_keys_str_mv | AT pittialexandre digitalcomputingthroughrandomnessandorderinneuralnetworks AT weidmannclaudio digitalcomputingthroughrandomnessandorderinneuralnetworks AT quoymathias digitalcomputingthroughrandomnessandorderinneuralnetworks |