Cargando…
An improved butterfly optimization algorithm for training the feed-forward artificial neural networks
Artificial neural network (ANN) which is an information processing technique developed by modeling the nervous system of the human brain is one of the most powerful learning methods today. One of the factors that make ANN successful is its training algorithm. In this paper, an improved butterfly opt...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Springer Berlin Heidelberg
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9584244/ https://www.ncbi.nlm.nih.gov/pubmed/36284902 http://dx.doi.org/10.1007/s00500-022-07592-w |
_version_ | 1784813221079678976 |
---|---|
author | Irmak, Büşra Karakoyun, Murat Gülcü, Şaban |
author_facet | Irmak, Büşra Karakoyun, Murat Gülcü, Şaban |
author_sort | Irmak, Büşra |
collection | PubMed |
description | Artificial neural network (ANN) which is an information processing technique developed by modeling the nervous system of the human brain is one of the most powerful learning methods today. One of the factors that make ANN successful is its training algorithm. In this paper, an improved butterfly optimization algorithm (IBOA) based on the butterfly optimization algorithm was proposed for training the feed-forward artificial neural networks. The IBOA algorithm has the chaotic property which helps optimization algorithms to explore the search space more dynamically and globally. In the experiments, ten chaotic maps were used. The success of the IBOA algorithm was tested on 13 benchmark functions which are well known to those working on global optimization and are frequently used for testing and analysis of optimization algorithms. The Tent-mapped IBOA algorithm outperformed the other algorithms in most of the benchmark functions. Moreover, the success of the IBOA-MLP algorithm also has been tested on five classification datasets (xor, balloon, iris, breast cancer, and heart) and the IBOA-MLP algorithm was compared with four algorithms in the literature. According to the statistical performance metrics (sensitivity, specificity, precision, F1-score, and Friedman test), the IBOA-MLP outperformed the other algorithms and proved to be successful in training the feed-forward artificial neural networks. |
format | Online Article Text |
id | pubmed-9584244 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Springer Berlin Heidelberg |
record_format | MEDLINE/PubMed |
spelling | pubmed-95842442022-10-21 An improved butterfly optimization algorithm for training the feed-forward artificial neural networks Irmak, Büşra Karakoyun, Murat Gülcü, Şaban Soft comput Optimization Artificial neural network (ANN) which is an information processing technique developed by modeling the nervous system of the human brain is one of the most powerful learning methods today. One of the factors that make ANN successful is its training algorithm. In this paper, an improved butterfly optimization algorithm (IBOA) based on the butterfly optimization algorithm was proposed for training the feed-forward artificial neural networks. The IBOA algorithm has the chaotic property which helps optimization algorithms to explore the search space more dynamically and globally. In the experiments, ten chaotic maps were used. The success of the IBOA algorithm was tested on 13 benchmark functions which are well known to those working on global optimization and are frequently used for testing and analysis of optimization algorithms. The Tent-mapped IBOA algorithm outperformed the other algorithms in most of the benchmark functions. Moreover, the success of the IBOA-MLP algorithm also has been tested on five classification datasets (xor, balloon, iris, breast cancer, and heart) and the IBOA-MLP algorithm was compared with four algorithms in the literature. According to the statistical performance metrics (sensitivity, specificity, precision, F1-score, and Friedman test), the IBOA-MLP outperformed the other algorithms and proved to be successful in training the feed-forward artificial neural networks. Springer Berlin Heidelberg 2022-10-20 2023 /pmc/articles/PMC9584244/ /pubmed/36284902 http://dx.doi.org/10.1007/s00500-022-07592-w Text en © The Author(s), under exclusive licence to Springer-Verlag GmbH Germany, part of Springer Nature 2022, Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law. This article is made available via the PMC Open Access Subset for unrestricted research re-use and secondary analysis in any form or by any means with acknowledgement of the original source. These permissions are granted for the duration of the World Health Organization (WHO) declaration of COVID-19 as a global pandemic. |
spellingShingle | Optimization Irmak, Büşra Karakoyun, Murat Gülcü, Şaban An improved butterfly optimization algorithm for training the feed-forward artificial neural networks |
title | An improved butterfly optimization algorithm for training the feed-forward artificial neural networks |
title_full | An improved butterfly optimization algorithm for training the feed-forward artificial neural networks |
title_fullStr | An improved butterfly optimization algorithm for training the feed-forward artificial neural networks |
title_full_unstemmed | An improved butterfly optimization algorithm for training the feed-forward artificial neural networks |
title_short | An improved butterfly optimization algorithm for training the feed-forward artificial neural networks |
title_sort | improved butterfly optimization algorithm for training the feed-forward artificial neural networks |
topic | Optimization |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9584244/ https://www.ncbi.nlm.nih.gov/pubmed/36284902 http://dx.doi.org/10.1007/s00500-022-07592-w |
work_keys_str_mv | AT irmakbusra animprovedbutterflyoptimizationalgorithmfortrainingthefeedforwardartificialneuralnetworks AT karakoyunmurat animprovedbutterflyoptimizationalgorithmfortrainingthefeedforwardartificialneuralnetworks AT gulcusaban animprovedbutterflyoptimizationalgorithmfortrainingthefeedforwardartificialneuralnetworks AT irmakbusra improvedbutterflyoptimizationalgorithmfortrainingthefeedforwardartificialneuralnetworks AT karakoyunmurat improvedbutterflyoptimizationalgorithmfortrainingthefeedforwardartificialneuralnetworks AT gulcusaban improvedbutterflyoptimizationalgorithmfortrainingthefeedforwardartificialneuralnetworks |