Cargando…

Multilayer perceptron architecture optimization using parallel computing techniques

The objective of this research was to develop a methodology for optimizing multilayer-perceptron-type neural networks by evaluating the effects of three neural architecture parameters, namely, number of hidden layers (HL), neurons per hidden layer (NHL), and activation function type (AF), on the sum...

Descripción completa

Detalles Bibliográficos
Autores principales: Castro, Wilson, Oblitas, Jimy, Santa-Cruz, Roberto, Avila-George, Himer
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Public Library of Science 2017
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5728525/
https://www.ncbi.nlm.nih.gov/pubmed/29236744
http://dx.doi.org/10.1371/journal.pone.0189369
_version_ 1783286039820894208
author Castro, Wilson
Oblitas, Jimy
Santa-Cruz, Roberto
Avila-George, Himer
author_facet Castro, Wilson
Oblitas, Jimy
Santa-Cruz, Roberto
Avila-George, Himer
author_sort Castro, Wilson
collection PubMed
description The objective of this research was to develop a methodology for optimizing multilayer-perceptron-type neural networks by evaluating the effects of three neural architecture parameters, namely, number of hidden layers (HL), neurons per hidden layer (NHL), and activation function type (AF), on the sum of squares error (SSE). The data for the study were obtained from quality parameters (physicochemical and microbiological) of milk samples. Architectures or combinations were organized in groups (G1, G2, and G3) generated upon interspersing one, two, and three layers. Within each group, the networks had three neurons in the input layer, six neurons in the output layer, three to twenty-seven NHL, and three AF (tan-sig, log-sig, and linear) types. The number of architectures was determined using three factorial-type experimental designs, which reached 63, 2 187, and 50 049 combinations for G1, G2 and G3, respectively. Using MATLAB 2015a, a logical sequence was designed and implemented for constructing, training, and evaluating multilayer-perceptron-type neural networks using parallel computing techniques. The results show that HL and NHL have a statistically relevant effect on SSE, and from two hidden layers, AF also has a significant effect; thus, both AF and NHL can be evaluated to determine the optimal combination per group. Moreover, in the three study groups, it is observed that there is an inverse relationship between the number of processors and the total optimization time.
format Online
Article
Text
id pubmed-5728525
institution National Center for Biotechnology Information
language English
publishDate 2017
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-57285252017-12-22 Multilayer perceptron architecture optimization using parallel computing techniques Castro, Wilson Oblitas, Jimy Santa-Cruz, Roberto Avila-George, Himer PLoS One Research Article The objective of this research was to develop a methodology for optimizing multilayer-perceptron-type neural networks by evaluating the effects of three neural architecture parameters, namely, number of hidden layers (HL), neurons per hidden layer (NHL), and activation function type (AF), on the sum of squares error (SSE). The data for the study were obtained from quality parameters (physicochemical and microbiological) of milk samples. Architectures or combinations were organized in groups (G1, G2, and G3) generated upon interspersing one, two, and three layers. Within each group, the networks had three neurons in the input layer, six neurons in the output layer, three to twenty-seven NHL, and three AF (tan-sig, log-sig, and linear) types. The number of architectures was determined using three factorial-type experimental designs, which reached 63, 2 187, and 50 049 combinations for G1, G2 and G3, respectively. Using MATLAB 2015a, a logical sequence was designed and implemented for constructing, training, and evaluating multilayer-perceptron-type neural networks using parallel computing techniques. The results show that HL and NHL have a statistically relevant effect on SSE, and from two hidden layers, AF also has a significant effect; thus, both AF and NHL can be evaluated to determine the optimal combination per group. Moreover, in the three study groups, it is observed that there is an inverse relationship between the number of processors and the total optimization time. Public Library of Science 2017-12-13 /pmc/articles/PMC5728525/ /pubmed/29236744 http://dx.doi.org/10.1371/journal.pone.0189369 Text en © 2017 Castro et al http://creativecommons.org/licenses/by/4.0/ This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
spellingShingle Research Article
Castro, Wilson
Oblitas, Jimy
Santa-Cruz, Roberto
Avila-George, Himer
Multilayer perceptron architecture optimization using parallel computing techniques
title Multilayer perceptron architecture optimization using parallel computing techniques
title_full Multilayer perceptron architecture optimization using parallel computing techniques
title_fullStr Multilayer perceptron architecture optimization using parallel computing techniques
title_full_unstemmed Multilayer perceptron architecture optimization using parallel computing techniques
title_short Multilayer perceptron architecture optimization using parallel computing techniques
title_sort multilayer perceptron architecture optimization using parallel computing techniques
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5728525/
https://www.ncbi.nlm.nih.gov/pubmed/29236744
http://dx.doi.org/10.1371/journal.pone.0189369
work_keys_str_mv AT castrowilson multilayerperceptronarchitectureoptimizationusingparallelcomputingtechniques
AT oblitasjimy multilayerperceptronarchitectureoptimizationusingparallelcomputingtechniques
AT santacruzroberto multilayerperceptronarchitectureoptimizationusingparallelcomputingtechniques
AT avilageorgehimer multilayerperceptronarchitectureoptimizationusingparallelcomputingtechniques