Cargando…
Recognition ability of untrained neural networks to symbolic numbers
Although animals can learn to use abstract numbers to represent the number of items, whether untrained animals could distinguish between different abstract numbers is not clear. A two-layer spiking neural network with lateral inhibition was built from the perspective of biological interpretability....
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9534535/ https://www.ncbi.nlm.nih.gov/pubmed/36213547 http://dx.doi.org/10.3389/fninf.2022.973010 |
_version_ | 1784802563925737472 |
---|---|
author | Zhou, Yiwei Chen, Huanwen Wang, Yijun |
author_facet | Zhou, Yiwei Chen, Huanwen Wang, Yijun |
author_sort | Zhou, Yiwei |
collection | PubMed |
description | Although animals can learn to use abstract numbers to represent the number of items, whether untrained animals could distinguish between different abstract numbers is not clear. A two-layer spiking neural network with lateral inhibition was built from the perspective of biological interpretability. The network connection weight was set randomly without adjustment. On the basis of this model, experiments were carried out on the symbolic number dataset MNIST and non-symbolic numerosity dataset. Results showed that the model has abilities to distinguish symbolic numbers. However, compared with number sense, tuning curves of symbolic numbers could not reproduce size and distance effects. The preference distribution also could not show high distribution characteristics at both ends and low distribution characteristics in the middle. More than half of the network units prefer the symbolic numbers 0 and 5. The average goodness-of-fit of the Gaussian fitting of tuning curves increases with the increase in abscissa non-linearity. These results revealed that the concept of human symbolic number is trained on the basis of number sense. |
format | Online Article Text |
id | pubmed-9534535 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-95345352022-10-06 Recognition ability of untrained neural networks to symbolic numbers Zhou, Yiwei Chen, Huanwen Wang, Yijun Front Neuroinform Neuroscience Although animals can learn to use abstract numbers to represent the number of items, whether untrained animals could distinguish between different abstract numbers is not clear. A two-layer spiking neural network with lateral inhibition was built from the perspective of biological interpretability. The network connection weight was set randomly without adjustment. On the basis of this model, experiments were carried out on the symbolic number dataset MNIST and non-symbolic numerosity dataset. Results showed that the model has abilities to distinguish symbolic numbers. However, compared with number sense, tuning curves of symbolic numbers could not reproduce size and distance effects. The preference distribution also could not show high distribution characteristics at both ends and low distribution characteristics in the middle. More than half of the network units prefer the symbolic numbers 0 and 5. The average goodness-of-fit of the Gaussian fitting of tuning curves increases with the increase in abscissa non-linearity. These results revealed that the concept of human symbolic number is trained on the basis of number sense. Frontiers Media S.A. 2022-09-21 /pmc/articles/PMC9534535/ /pubmed/36213547 http://dx.doi.org/10.3389/fninf.2022.973010 Text en Copyright © 2022 Zhou, Chen and Wang. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Neuroscience Zhou, Yiwei Chen, Huanwen Wang, Yijun Recognition ability of untrained neural networks to symbolic numbers |
title | Recognition ability of untrained neural networks to symbolic numbers |
title_full | Recognition ability of untrained neural networks to symbolic numbers |
title_fullStr | Recognition ability of untrained neural networks to symbolic numbers |
title_full_unstemmed | Recognition ability of untrained neural networks to symbolic numbers |
title_short | Recognition ability of untrained neural networks to symbolic numbers |
title_sort | recognition ability of untrained neural networks to symbolic numbers |
topic | Neuroscience |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9534535/ https://www.ncbi.nlm.nih.gov/pubmed/36213547 http://dx.doi.org/10.3389/fninf.2022.973010 |
work_keys_str_mv | AT zhouyiwei recognitionabilityofuntrainedneuralnetworkstosymbolicnumbers AT chenhuanwen recognitionabilityofuntrainedneuralnetworkstosymbolicnumbers AT wangyijun recognitionabilityofuntrainedneuralnetworkstosymbolicnumbers |