Cargando…

Understanding the Impact of Neural Variations and Random Connections on Inference

Recent research suggests that in vitro neural networks created from dissociated neurons may be used for computing and performing machine learning tasks. To develop a better artificial intelligent system, a hybrid bio-silicon computer is worth exploring, but its performance is still inferior to that...

Descripción completa

Detalles Bibliográficos
Autores principales: Zeng, Yuan, Ferdous, Zubayer Ibne, Zhang, Weixiang, Xu, Mufan, Yu, Anlan, Patel, Drew, Post, Valentin, Guo, Xiaochen, Berdichevsky, Yevgeny, Yan, Zhiyuan
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8215547/
https://www.ncbi.nlm.nih.gov/pubmed/34163343
http://dx.doi.org/10.3389/fncom.2021.612937
_version_ 1783710264771739648
author Zeng, Yuan
Ferdous, Zubayer Ibne
Zhang, Weixiang
Xu, Mufan
Yu, Anlan
Patel, Drew
Post, Valentin
Guo, Xiaochen
Berdichevsky, Yevgeny
Yan, Zhiyuan
author_facet Zeng, Yuan
Ferdous, Zubayer Ibne
Zhang, Weixiang
Xu, Mufan
Yu, Anlan
Patel, Drew
Post, Valentin
Guo, Xiaochen
Berdichevsky, Yevgeny
Yan, Zhiyuan
author_sort Zeng, Yuan
collection PubMed
description Recent research suggests that in vitro neural networks created from dissociated neurons may be used for computing and performing machine learning tasks. To develop a better artificial intelligent system, a hybrid bio-silicon computer is worth exploring, but its performance is still inferior to that of a silicon-based computer. One reason may be that a living neural network has many intrinsic properties, such as random network connectivity, high network sparsity, and large neural and synaptic variability. These properties may lead to new design considerations, and existing algorithms need to be adjusted for living neural network implementation. This work investigates the impact of neural variations and random connections on inference with learning algorithms. A two-layer hybrid bio-silicon platform is constructed and a five-step design method is proposed for the fast development of living neural network algorithms. Neural variations and dynamics are verified by fitting model parameters with biological experimental results. Random connections are generated under different connection probabilities to vary network sparsity. A multi-layer perceptron algorithm is tested with biological constraints on the MNIST dataset. The results show that a reasonable inference accuracy can be achieved despite the presence of neural variations and random network connections. A new adaptive pre-processing technique is proposed to ensure good learning accuracy with different living neural network sparsity.
format Online
Article
Text
id pubmed-8215547
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-82155472021-06-22 Understanding the Impact of Neural Variations and Random Connections on Inference Zeng, Yuan Ferdous, Zubayer Ibne Zhang, Weixiang Xu, Mufan Yu, Anlan Patel, Drew Post, Valentin Guo, Xiaochen Berdichevsky, Yevgeny Yan, Zhiyuan Front Comput Neurosci Neuroscience Recent research suggests that in vitro neural networks created from dissociated neurons may be used for computing and performing machine learning tasks. To develop a better artificial intelligent system, a hybrid bio-silicon computer is worth exploring, but its performance is still inferior to that of a silicon-based computer. One reason may be that a living neural network has many intrinsic properties, such as random network connectivity, high network sparsity, and large neural and synaptic variability. These properties may lead to new design considerations, and existing algorithms need to be adjusted for living neural network implementation. This work investigates the impact of neural variations and random connections on inference with learning algorithms. A two-layer hybrid bio-silicon platform is constructed and a five-step design method is proposed for the fast development of living neural network algorithms. Neural variations and dynamics are verified by fitting model parameters with biological experimental results. Random connections are generated under different connection probabilities to vary network sparsity. A multi-layer perceptron algorithm is tested with biological constraints on the MNIST dataset. The results show that a reasonable inference accuracy can be achieved despite the presence of neural variations and random network connections. A new adaptive pre-processing technique is proposed to ensure good learning accuracy with different living neural network sparsity. Frontiers Media S.A. 2021-06-07 /pmc/articles/PMC8215547/ /pubmed/34163343 http://dx.doi.org/10.3389/fncom.2021.612937 Text en Copyright © 2021 Zeng, Ferdous, Zhang, Xu, Yu, Patel, Post, Guo, Berdichevsky and Yan. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Neuroscience
Zeng, Yuan
Ferdous, Zubayer Ibne
Zhang, Weixiang
Xu, Mufan
Yu, Anlan
Patel, Drew
Post, Valentin
Guo, Xiaochen
Berdichevsky, Yevgeny
Yan, Zhiyuan
Understanding the Impact of Neural Variations and Random Connections on Inference
title Understanding the Impact of Neural Variations and Random Connections on Inference
title_full Understanding the Impact of Neural Variations and Random Connections on Inference
title_fullStr Understanding the Impact of Neural Variations and Random Connections on Inference
title_full_unstemmed Understanding the Impact of Neural Variations and Random Connections on Inference
title_short Understanding the Impact of Neural Variations and Random Connections on Inference
title_sort understanding the impact of neural variations and random connections on inference
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8215547/
https://www.ncbi.nlm.nih.gov/pubmed/34163343
http://dx.doi.org/10.3389/fncom.2021.612937
work_keys_str_mv AT zengyuan understandingtheimpactofneuralvariationsandrandomconnectionsoninference
AT ferdouszubayeribne understandingtheimpactofneuralvariationsandrandomconnectionsoninference
AT zhangweixiang understandingtheimpactofneuralvariationsandrandomconnectionsoninference
AT xumufan understandingtheimpactofneuralvariationsandrandomconnectionsoninference
AT yuanlan understandingtheimpactofneuralvariationsandrandomconnectionsoninference
AT pateldrew understandingtheimpactofneuralvariationsandrandomconnectionsoninference
AT postvalentin understandingtheimpactofneuralvariationsandrandomconnectionsoninference
AT guoxiaochen understandingtheimpactofneuralvariationsandrandomconnectionsoninference
AT berdichevskyyevgeny understandingtheimpactofneuralvariationsandrandomconnectionsoninference
AT yanzhiyuan understandingtheimpactofneuralvariationsandrandomconnectionsoninference