Cargando…

Multi-objective simulated annealing for hyper-parameter optimization in convolutional neural networks

In this study, we model a CNN hyper-parameter optimization problem as a bi-criteria optimization problem, where the first objective being the classification accuracy and the second objective being the computational complexity which is measured in terms of the number of floating point operations. For...

Descripción completa

Detalles Bibliográficos
Autores principales: Gülcü, Ayla, Kuş, Zeki
Formato: Online Artículo Texto
Lenguaje:English
Publicado: PeerJ Inc. 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7924536/
https://www.ncbi.nlm.nih.gov/pubmed/33816989
http://dx.doi.org/10.7717/peerj-cs.338
_version_ 1783659110195003392
author Gülcü, Ayla
Kuş, Zeki
author_facet Gülcü, Ayla
Kuş, Zeki
author_sort Gülcü, Ayla
collection PubMed
description In this study, we model a CNN hyper-parameter optimization problem as a bi-criteria optimization problem, where the first objective being the classification accuracy and the second objective being the computational complexity which is measured in terms of the number of floating point operations. For this bi-criteria optimization problem, we develop a Multi-Objective Simulated Annealing (MOSA) algorithm for obtaining high-quality solutions in terms of both objectives. CIFAR-10 is selected as the benchmark dataset, and the MOSA trade-off fronts obtained for this dataset are compared to the fronts generated by a single-objective Simulated Annealing (SA) algorithm with respect to several front evaluation metrics such as generational distance, spacing and spread. The comparison results suggest that the MOSA algorithm is able to search the objective space more effectively than the SA method. For each of these methods, some front solutions are selected for longer training in order to see their actual performance on the original test set. Again, the results state that the MOSA performs better than the SA under multi-objective setting. The performance of the MOSA configurations are also compared to other search generated and human designed state-of-the-art architectures. It is shown that the network configurations generated by the MOSA are not dominated by those architectures, and the proposed method can be of great use when the computational complexity is as important as the test accuracy.
format Online
Article
Text
id pubmed-7924536
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher PeerJ Inc.
record_format MEDLINE/PubMed
spelling pubmed-79245362021-04-02 Multi-objective simulated annealing for hyper-parameter optimization in convolutional neural networks Gülcü, Ayla Kuş, Zeki PeerJ Comput Sci Artificial Intelligence In this study, we model a CNN hyper-parameter optimization problem as a bi-criteria optimization problem, where the first objective being the classification accuracy and the second objective being the computational complexity which is measured in terms of the number of floating point operations. For this bi-criteria optimization problem, we develop a Multi-Objective Simulated Annealing (MOSA) algorithm for obtaining high-quality solutions in terms of both objectives. CIFAR-10 is selected as the benchmark dataset, and the MOSA trade-off fronts obtained for this dataset are compared to the fronts generated by a single-objective Simulated Annealing (SA) algorithm with respect to several front evaluation metrics such as generational distance, spacing and spread. The comparison results suggest that the MOSA algorithm is able to search the objective space more effectively than the SA method. For each of these methods, some front solutions are selected for longer training in order to see their actual performance on the original test set. Again, the results state that the MOSA performs better than the SA under multi-objective setting. The performance of the MOSA configurations are also compared to other search generated and human designed state-of-the-art architectures. It is shown that the network configurations generated by the MOSA are not dominated by those architectures, and the proposed method can be of great use when the computational complexity is as important as the test accuracy. PeerJ Inc. 2021-01-04 /pmc/articles/PMC7924536/ /pubmed/33816989 http://dx.doi.org/10.7717/peerj-cs.338 Text en © 2021 Gülcü and Kuş https://creativecommons.org/licenses/by/4.0/ This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, reproduction and adaptation in any medium and for any purpose provided that it is properly attributed. For attribution, the original author(s), title, publication source (PeerJ Computer Science) and either DOI or URL of the article must be cited.
spellingShingle Artificial Intelligence
Gülcü, Ayla
Kuş, Zeki
Multi-objective simulated annealing for hyper-parameter optimization in convolutional neural networks
title Multi-objective simulated annealing for hyper-parameter optimization in convolutional neural networks
title_full Multi-objective simulated annealing for hyper-parameter optimization in convolutional neural networks
title_fullStr Multi-objective simulated annealing for hyper-parameter optimization in convolutional neural networks
title_full_unstemmed Multi-objective simulated annealing for hyper-parameter optimization in convolutional neural networks
title_short Multi-objective simulated annealing for hyper-parameter optimization in convolutional neural networks
title_sort multi-objective simulated annealing for hyper-parameter optimization in convolutional neural networks
topic Artificial Intelligence
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7924536/
https://www.ncbi.nlm.nih.gov/pubmed/33816989
http://dx.doi.org/10.7717/peerj-cs.338
work_keys_str_mv AT gulcuayla multiobjectivesimulatedannealingforhyperparameteroptimizationinconvolutionalneuralnetworks
AT kuszeki multiobjectivesimulatedannealingforhyperparameteroptimizationinconvolutionalneuralnetworks