Cargando…
Less Is More: Adaptive Trainable Gradient Dropout for Deep Neural Networks
The undeniable computational power of artificial neural networks has granted the scientific community the ability to exploit the available data in ways previously inconceivable. However, deep neural networks require an overwhelming quantity of data in order to interpret the underlying connections be...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9921448/ https://www.ncbi.nlm.nih.gov/pubmed/36772365 http://dx.doi.org/10.3390/s23031325 |
_version_ | 1784887314776850432 |
---|---|
author | Avgerinos, Christos Vretos, Nicholas Daras, Petros |
author_facet | Avgerinos, Christos Vretos, Nicholas Daras, Petros |
author_sort | Avgerinos, Christos |
collection | PubMed |
description | The undeniable computational power of artificial neural networks has granted the scientific community the ability to exploit the available data in ways previously inconceivable. However, deep neural networks require an overwhelming quantity of data in order to interpret the underlying connections between them, and therefore, be able to complete the specific task that they have been assigned to. Feeding a deep neural network with vast amounts of data usually ensures efficiency, but may, however, harm the network’s ability to generalize. To tackle this, numerous regularization techniques have been proposed, with dropout being one of the most dominant. This paper proposes a selective gradient dropout method, which, instead of relying on dropping random weights, learns to freeze the training process of specific connections, thereby increasing the overall network’s sparsity in an adaptive manner, by driving it to utilize more salient weights. The experimental results show that the produced sparse network outperforms the baseline on numerous image classification datasets, and additionally, the yielded results occurred after significantly less training epochs. |
format | Online Article Text |
id | pubmed-9921448 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-99214482023-02-12 Less Is More: Adaptive Trainable Gradient Dropout for Deep Neural Networks Avgerinos, Christos Vretos, Nicholas Daras, Petros Sensors (Basel) Article The undeniable computational power of artificial neural networks has granted the scientific community the ability to exploit the available data in ways previously inconceivable. However, deep neural networks require an overwhelming quantity of data in order to interpret the underlying connections between them, and therefore, be able to complete the specific task that they have been assigned to. Feeding a deep neural network with vast amounts of data usually ensures efficiency, but may, however, harm the network’s ability to generalize. To tackle this, numerous regularization techniques have been proposed, with dropout being one of the most dominant. This paper proposes a selective gradient dropout method, which, instead of relying on dropping random weights, learns to freeze the training process of specific connections, thereby increasing the overall network’s sparsity in an adaptive manner, by driving it to utilize more salient weights. The experimental results show that the produced sparse network outperforms the baseline on numerous image classification datasets, and additionally, the yielded results occurred after significantly less training epochs. MDPI 2023-01-24 /pmc/articles/PMC9921448/ /pubmed/36772365 http://dx.doi.org/10.3390/s23031325 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Avgerinos, Christos Vretos, Nicholas Daras, Petros Less Is More: Adaptive Trainable Gradient Dropout for Deep Neural Networks |
title | Less Is More: Adaptive Trainable Gradient Dropout for Deep Neural Networks |
title_full | Less Is More: Adaptive Trainable Gradient Dropout for Deep Neural Networks |
title_fullStr | Less Is More: Adaptive Trainable Gradient Dropout for Deep Neural Networks |
title_full_unstemmed | Less Is More: Adaptive Trainable Gradient Dropout for Deep Neural Networks |
title_short | Less Is More: Adaptive Trainable Gradient Dropout for Deep Neural Networks |
title_sort | less is more: adaptive trainable gradient dropout for deep neural networks |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9921448/ https://www.ncbi.nlm.nih.gov/pubmed/36772365 http://dx.doi.org/10.3390/s23031325 |
work_keys_str_mv | AT avgerinoschristos lessismoreadaptivetrainablegradientdropoutfordeepneuralnetworks AT vretosnicholas lessismoreadaptivetrainablegradientdropoutfordeepneuralnetworks AT daraspetros lessismoreadaptivetrainablegradientdropoutfordeepneuralnetworks |