Cargando…

Comparison of Different Convolutional Neural Network Activation Functions and Methods for Building Ensembles for Small to Midsize Medical Data Sets

CNNs and other deep learners are now state-of-the-art in medical imaging research. However, the small sample size of many medical data sets dampens performance and results in overfitting. In some medical areas, it is simply too labor-intensive and expensive to amass images numbering in the hundreds...

Descripción completa

Detalles Bibliográficos
Autores principales: Nanni, Loris, Brahnam, Sheryl, Paci, Michelangelo, Ghidoni, Stefano
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9415767/
https://www.ncbi.nlm.nih.gov/pubmed/36015898
http://dx.doi.org/10.3390/s22166129
_version_ 1784776313452625920
author Nanni, Loris
Brahnam, Sheryl
Paci, Michelangelo
Ghidoni, Stefano
author_facet Nanni, Loris
Brahnam, Sheryl
Paci, Michelangelo
Ghidoni, Stefano
author_sort Nanni, Loris
collection PubMed
description CNNs and other deep learners are now state-of-the-art in medical imaging research. However, the small sample size of many medical data sets dampens performance and results in overfitting. In some medical areas, it is simply too labor-intensive and expensive to amass images numbering in the hundreds of thousands. Building Deep CNN ensembles of pre-trained CNNs is one powerful method for overcoming this problem. Ensembles combine the outputs of multiple classifiers to improve performance. This method relies on the introduction of diversity, which can be introduced on many levels in the classification workflow. A recent ensembling method that has shown promise is to vary the activation functions in a set of CNNs or within different layers of a single CNN. This study aims to examine the performance of both methods using a large set of twenty activations functions, six of which are presented here for the first time: 2D Mexican ReLU, TanELU, MeLU + GaLU, Symmetric MeLU, Symmetric GaLU, and Flexible MeLU. The proposed method was tested on fifteen medical data sets representing various classification tasks. The best performing ensemble combined two well-known CNNs (VGG16 and ResNet50) whose standard ReLU activation layers were randomly replaced with another. Results demonstrate the superiority in performance of this approach.
format Online
Article
Text
id pubmed-9415767
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-94157672022-08-27 Comparison of Different Convolutional Neural Network Activation Functions and Methods for Building Ensembles for Small to Midsize Medical Data Sets Nanni, Loris Brahnam, Sheryl Paci, Michelangelo Ghidoni, Stefano Sensors (Basel) Article CNNs and other deep learners are now state-of-the-art in medical imaging research. However, the small sample size of many medical data sets dampens performance and results in overfitting. In some medical areas, it is simply too labor-intensive and expensive to amass images numbering in the hundreds of thousands. Building Deep CNN ensembles of pre-trained CNNs is one powerful method for overcoming this problem. Ensembles combine the outputs of multiple classifiers to improve performance. This method relies on the introduction of diversity, which can be introduced on many levels in the classification workflow. A recent ensembling method that has shown promise is to vary the activation functions in a set of CNNs or within different layers of a single CNN. This study aims to examine the performance of both methods using a large set of twenty activations functions, six of which are presented here for the first time: 2D Mexican ReLU, TanELU, MeLU + GaLU, Symmetric MeLU, Symmetric GaLU, and Flexible MeLU. The proposed method was tested on fifteen medical data sets representing various classification tasks. The best performing ensemble combined two well-known CNNs (VGG16 and ResNet50) whose standard ReLU activation layers were randomly replaced with another. Results demonstrate the superiority in performance of this approach. MDPI 2022-08-16 /pmc/articles/PMC9415767/ /pubmed/36015898 http://dx.doi.org/10.3390/s22166129 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Nanni, Loris
Brahnam, Sheryl
Paci, Michelangelo
Ghidoni, Stefano
Comparison of Different Convolutional Neural Network Activation Functions and Methods for Building Ensembles for Small to Midsize Medical Data Sets
title Comparison of Different Convolutional Neural Network Activation Functions and Methods for Building Ensembles for Small to Midsize Medical Data Sets
title_full Comparison of Different Convolutional Neural Network Activation Functions and Methods for Building Ensembles for Small to Midsize Medical Data Sets
title_fullStr Comparison of Different Convolutional Neural Network Activation Functions and Methods for Building Ensembles for Small to Midsize Medical Data Sets
title_full_unstemmed Comparison of Different Convolutional Neural Network Activation Functions and Methods for Building Ensembles for Small to Midsize Medical Data Sets
title_short Comparison of Different Convolutional Neural Network Activation Functions and Methods for Building Ensembles for Small to Midsize Medical Data Sets
title_sort comparison of different convolutional neural network activation functions and methods for building ensembles for small to midsize medical data sets
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9415767/
https://www.ncbi.nlm.nih.gov/pubmed/36015898
http://dx.doi.org/10.3390/s22166129
work_keys_str_mv AT nanniloris comparisonofdifferentconvolutionalneuralnetworkactivationfunctionsandmethodsforbuildingensemblesforsmalltomidsizemedicaldatasets
AT brahnamsheryl comparisonofdifferentconvolutionalneuralnetworkactivationfunctionsandmethodsforbuildingensemblesforsmalltomidsizemedicaldatasets
AT pacimichelangelo comparisonofdifferentconvolutionalneuralnetworkactivationfunctionsandmethodsforbuildingensemblesforsmalltomidsizemedicaldatasets
AT ghidonistefano comparisonofdifferentconvolutionalneuralnetworkactivationfunctionsandmethodsforbuildingensemblesforsmalltomidsizemedicaldatasets