Cargando…

Stochastic Selection of Activation Layers for Convolutional Neural Networks

In recent years, the field of deep learning has achieved considerable success in pattern recognition, image segmentation, and many other classification fields. There are many studies and practical applications of deep learning on images, video, or text classification. Activation functions play a cru...

Descripción completa

Detalles Bibliográficos
Autores principales: Nanni, Loris, Lumini, Alessandra, Ghidoni, Stefano, Maguolo, Gianluca
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7147370/
https://www.ncbi.nlm.nih.gov/pubmed/32183334
http://dx.doi.org/10.3390/s20061626
_version_ 1783520406458597376
author Nanni, Loris
Lumini, Alessandra
Ghidoni, Stefano
Maguolo, Gianluca
author_facet Nanni, Loris
Lumini, Alessandra
Ghidoni, Stefano
Maguolo, Gianluca
author_sort Nanni, Loris
collection PubMed
description In recent years, the field of deep learning has achieved considerable success in pattern recognition, image segmentation, and many other classification fields. There are many studies and practical applications of deep learning on images, video, or text classification. Activation functions play a crucial role in discriminative capabilities of the deep neural networks and the design of new “static” or “dynamic” activation functions is an active area of research. The main difference between “static” and “dynamic” functions is that the first class of activations considers all the neurons and layers as identical, while the second class learns parameters of the activation function independently for each layer or even each neuron. Although the “dynamic” activation functions perform better in some applications, the increased number of trainable parameters requires more computational time and can lead to overfitting. In this work, we propose a mixture of “static” and “dynamic” activation functions, which are stochastically selected at each layer. Our idea for model design is based on a method for changing some layers along the lines of different functional blocks of the best performing CNN models, with the aim of designing new models to be used as stand-alone networks or as a component of an ensemble. We propose to replace each activation layer of a CNN (usually a ReLU layer) by a different activation function stochastically drawn from a set of activation functions: in this way, the resulting CNN has a different set of activation function layers.
format Online
Article
Text
id pubmed-7147370
institution National Center for Biotechnology Information
language English
publishDate 2020
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-71473702020-04-20 Stochastic Selection of Activation Layers for Convolutional Neural Networks Nanni, Loris Lumini, Alessandra Ghidoni, Stefano Maguolo, Gianluca Sensors (Basel) Article In recent years, the field of deep learning has achieved considerable success in pattern recognition, image segmentation, and many other classification fields. There are many studies and practical applications of deep learning on images, video, or text classification. Activation functions play a crucial role in discriminative capabilities of the deep neural networks and the design of new “static” or “dynamic” activation functions is an active area of research. The main difference between “static” and “dynamic” functions is that the first class of activations considers all the neurons and layers as identical, while the second class learns parameters of the activation function independently for each layer or even each neuron. Although the “dynamic” activation functions perform better in some applications, the increased number of trainable parameters requires more computational time and can lead to overfitting. In this work, we propose a mixture of “static” and “dynamic” activation functions, which are stochastically selected at each layer. Our idea for model design is based on a method for changing some layers along the lines of different functional blocks of the best performing CNN models, with the aim of designing new models to be used as stand-alone networks or as a component of an ensemble. We propose to replace each activation layer of a CNN (usually a ReLU layer) by a different activation function stochastically drawn from a set of activation functions: in this way, the resulting CNN has a different set of activation function layers. MDPI 2020-03-14 /pmc/articles/PMC7147370/ /pubmed/32183334 http://dx.doi.org/10.3390/s20061626 Text en © 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Nanni, Loris
Lumini, Alessandra
Ghidoni, Stefano
Maguolo, Gianluca
Stochastic Selection of Activation Layers for Convolutional Neural Networks
title Stochastic Selection of Activation Layers for Convolutional Neural Networks
title_full Stochastic Selection of Activation Layers for Convolutional Neural Networks
title_fullStr Stochastic Selection of Activation Layers for Convolutional Neural Networks
title_full_unstemmed Stochastic Selection of Activation Layers for Convolutional Neural Networks
title_short Stochastic Selection of Activation Layers for Convolutional Neural Networks
title_sort stochastic selection of activation layers for convolutional neural networks
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7147370/
https://www.ncbi.nlm.nih.gov/pubmed/32183334
http://dx.doi.org/10.3390/s20061626
work_keys_str_mv AT nanniloris stochasticselectionofactivationlayersforconvolutionalneuralnetworks
AT luminialessandra stochasticselectionofactivationlayersforconvolutionalneuralnetworks
AT ghidonistefano stochasticselectionofactivationlayersforconvolutionalneuralnetworks
AT maguologianluca stochasticselectionofactivationlayersforconvolutionalneuralnetworks