Cargando…
Learnability of the Boolean Innerproduct in Deep Neural Networks
In this paper, we study the learnability of the Boolean inner product by a systematic simulation study. The family of the Boolean inner product function is known to be representable by neural networks of threshold neurons of depth 3 with only [Formula: see text] units (n the input dimension)—whereas...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9407482/ https://www.ncbi.nlm.nih.gov/pubmed/36010780 http://dx.doi.org/10.3390/e24081117 |
_version_ | 1784774374789742592 |
---|---|
author | Erdal, Mehmet Schwenker, Friedhelm |
author_facet | Erdal, Mehmet Schwenker, Friedhelm |
author_sort | Erdal, Mehmet |
collection | PubMed |
description | In this paper, we study the learnability of the Boolean inner product by a systematic simulation study. The family of the Boolean inner product function is known to be representable by neural networks of threshold neurons of depth 3 with only [Formula: see text] units (n the input dimension)—whereas an exact representation by a depth 2 network cannot possibly be of polynomial size. This result can be seen as a strong argument for deep neural network architectures. In our study, we found that this depth 3 architecture of the Boolean inner product is difficult to train, much harder than the depth 2 network, at least for the small input size scenarios [Formula: see text]. Nonetheless, the accuracy of the deep architecture increased with the dimension of the input space to 94% on average, which means that multiple restarts are needed to find the compact depth 3 architecture. Replacing the fully connected first layer by a partially connected layer (a kind of convolutional layer sparsely connected with weight sharing) can significantly improve the learning performance up to 99% accuracy in simulations. Another way to improve the learnability of the compact depth 3 representation of the inner product could be achieved by adding just a few additional units into the first hidden layer. |
format | Online Article Text |
id | pubmed-9407482 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-94074822022-08-26 Learnability of the Boolean Innerproduct in Deep Neural Networks Erdal, Mehmet Schwenker, Friedhelm Entropy (Basel) Article In this paper, we study the learnability of the Boolean inner product by a systematic simulation study. The family of the Boolean inner product function is known to be representable by neural networks of threshold neurons of depth 3 with only [Formula: see text] units (n the input dimension)—whereas an exact representation by a depth 2 network cannot possibly be of polynomial size. This result can be seen as a strong argument for deep neural network architectures. In our study, we found that this depth 3 architecture of the Boolean inner product is difficult to train, much harder than the depth 2 network, at least for the small input size scenarios [Formula: see text]. Nonetheless, the accuracy of the deep architecture increased with the dimension of the input space to 94% on average, which means that multiple restarts are needed to find the compact depth 3 architecture. Replacing the fully connected first layer by a partially connected layer (a kind of convolutional layer sparsely connected with weight sharing) can significantly improve the learning performance up to 99% accuracy in simulations. Another way to improve the learnability of the compact depth 3 representation of the inner product could be achieved by adding just a few additional units into the first hidden layer. MDPI 2022-08-13 /pmc/articles/PMC9407482/ /pubmed/36010780 http://dx.doi.org/10.3390/e24081117 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Erdal, Mehmet Schwenker, Friedhelm Learnability of the Boolean Innerproduct in Deep Neural Networks |
title | Learnability of the Boolean Innerproduct in Deep Neural Networks |
title_full | Learnability of the Boolean Innerproduct in Deep Neural Networks |
title_fullStr | Learnability of the Boolean Innerproduct in Deep Neural Networks |
title_full_unstemmed | Learnability of the Boolean Innerproduct in Deep Neural Networks |
title_short | Learnability of the Boolean Innerproduct in Deep Neural Networks |
title_sort | learnability of the boolean innerproduct in deep neural networks |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9407482/ https://www.ncbi.nlm.nih.gov/pubmed/36010780 http://dx.doi.org/10.3390/e24081117 |
work_keys_str_mv | AT erdalmehmet learnabilityofthebooleaninnerproductindeepneuralnetworks AT schwenkerfriedhelm learnabilityofthebooleaninnerproductindeepneuralnetworks |