Cargando…
Comparison of Different Convolutional Neural Network Activation Functions and Methods for Building Ensembles for Small to Midsize Medical Data Sets
CNNs and other deep learners are now state-of-the-art in medical imaging research. However, the small sample size of many medical data sets dampens performance and results in overfitting. In some medical areas, it is simply too labor-intensive and expensive to amass images numbering in the hundreds...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9415767/ https://www.ncbi.nlm.nih.gov/pubmed/36015898 http://dx.doi.org/10.3390/s22166129 |
_version_ | 1784776313452625920 |
---|---|
author | Nanni, Loris Brahnam, Sheryl Paci, Michelangelo Ghidoni, Stefano |
author_facet | Nanni, Loris Brahnam, Sheryl Paci, Michelangelo Ghidoni, Stefano |
author_sort | Nanni, Loris |
collection | PubMed |
description | CNNs and other deep learners are now state-of-the-art in medical imaging research. However, the small sample size of many medical data sets dampens performance and results in overfitting. In some medical areas, it is simply too labor-intensive and expensive to amass images numbering in the hundreds of thousands. Building Deep CNN ensembles of pre-trained CNNs is one powerful method for overcoming this problem. Ensembles combine the outputs of multiple classifiers to improve performance. This method relies on the introduction of diversity, which can be introduced on many levels in the classification workflow. A recent ensembling method that has shown promise is to vary the activation functions in a set of CNNs or within different layers of a single CNN. This study aims to examine the performance of both methods using a large set of twenty activations functions, six of which are presented here for the first time: 2D Mexican ReLU, TanELU, MeLU + GaLU, Symmetric MeLU, Symmetric GaLU, and Flexible MeLU. The proposed method was tested on fifteen medical data sets representing various classification tasks. The best performing ensemble combined two well-known CNNs (VGG16 and ResNet50) whose standard ReLU activation layers were randomly replaced with another. Results demonstrate the superiority in performance of this approach. |
format | Online Article Text |
id | pubmed-9415767 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-94157672022-08-27 Comparison of Different Convolutional Neural Network Activation Functions and Methods for Building Ensembles for Small to Midsize Medical Data Sets Nanni, Loris Brahnam, Sheryl Paci, Michelangelo Ghidoni, Stefano Sensors (Basel) Article CNNs and other deep learners are now state-of-the-art in medical imaging research. However, the small sample size of many medical data sets dampens performance and results in overfitting. In some medical areas, it is simply too labor-intensive and expensive to amass images numbering in the hundreds of thousands. Building Deep CNN ensembles of pre-trained CNNs is one powerful method for overcoming this problem. Ensembles combine the outputs of multiple classifiers to improve performance. This method relies on the introduction of diversity, which can be introduced on many levels in the classification workflow. A recent ensembling method that has shown promise is to vary the activation functions in a set of CNNs or within different layers of a single CNN. This study aims to examine the performance of both methods using a large set of twenty activations functions, six of which are presented here for the first time: 2D Mexican ReLU, TanELU, MeLU + GaLU, Symmetric MeLU, Symmetric GaLU, and Flexible MeLU. The proposed method was tested on fifteen medical data sets representing various classification tasks. The best performing ensemble combined two well-known CNNs (VGG16 and ResNet50) whose standard ReLU activation layers were randomly replaced with another. Results demonstrate the superiority in performance of this approach. MDPI 2022-08-16 /pmc/articles/PMC9415767/ /pubmed/36015898 http://dx.doi.org/10.3390/s22166129 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Nanni, Loris Brahnam, Sheryl Paci, Michelangelo Ghidoni, Stefano Comparison of Different Convolutional Neural Network Activation Functions and Methods for Building Ensembles for Small to Midsize Medical Data Sets |
title | Comparison of Different Convolutional Neural Network Activation Functions and Methods for Building Ensembles for Small to Midsize Medical Data Sets |
title_full | Comparison of Different Convolutional Neural Network Activation Functions and Methods for Building Ensembles for Small to Midsize Medical Data Sets |
title_fullStr | Comparison of Different Convolutional Neural Network Activation Functions and Methods for Building Ensembles for Small to Midsize Medical Data Sets |
title_full_unstemmed | Comparison of Different Convolutional Neural Network Activation Functions and Methods for Building Ensembles for Small to Midsize Medical Data Sets |
title_short | Comparison of Different Convolutional Neural Network Activation Functions and Methods for Building Ensembles for Small to Midsize Medical Data Sets |
title_sort | comparison of different convolutional neural network activation functions and methods for building ensembles for small to midsize medical data sets |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9415767/ https://www.ncbi.nlm.nih.gov/pubmed/36015898 http://dx.doi.org/10.3390/s22166129 |
work_keys_str_mv | AT nanniloris comparisonofdifferentconvolutionalneuralnetworkactivationfunctionsandmethodsforbuildingensemblesforsmalltomidsizemedicaldatasets AT brahnamsheryl comparisonofdifferentconvolutionalneuralnetworkactivationfunctionsandmethodsforbuildingensemblesforsmalltomidsizemedicaldatasets AT pacimichelangelo comparisonofdifferentconvolutionalneuralnetworkactivationfunctionsandmethodsforbuildingensemblesforsmalltomidsizemedicaldatasets AT ghidonistefano comparisonofdifferentconvolutionalneuralnetworkactivationfunctionsandmethodsforbuildingensemblesforsmalltomidsizemedicaldatasets |