Cargando…

Study on the Use of Artificially Generated Objects in the Process of Training MLP Neural Networks Based on Dispersed Data

This study concerns dispersed data stored in independent local tables with different sets of attributes. The paper proposes a new method for training a single neural network—a multilayer perceptron based on dispersed data. The idea is to train local models that have identical structures based on loc...

Descripción completa

Detalles Bibliográficos
Autores principales: Marfo, Kwabena Frimpong, Przybyła-Kasperek, Małgorzata
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10217427/
https://www.ncbi.nlm.nih.gov/pubmed/37238459
http://dx.doi.org/10.3390/e25050703
_version_ 1785048535249453056
author Marfo, Kwabena Frimpong
Przybyła-Kasperek, Małgorzata
author_facet Marfo, Kwabena Frimpong
Przybyła-Kasperek, Małgorzata
author_sort Marfo, Kwabena Frimpong
collection PubMed
description This study concerns dispersed data stored in independent local tables with different sets of attributes. The paper proposes a new method for training a single neural network—a multilayer perceptron based on dispersed data. The idea is to train local models that have identical structures based on local tables; however, due to different sets of conditional attributes present in local tables, it is necessary to generate some artificial objects to train local models. The paper presents a study on the use of varying parameter values in the proposed method of creating artificial objects to train local models. The paper presents an exhaustive comparison in terms of the number of artificial objects generated based on a single original object, the degree of data dispersion, data balancing, and different network structures—the number of neurons in the hidden layer. It was found that for data sets with a large number of objects, a smaller number of artificial objects is optimal. For smaller data sets, a greater number of artificial objects (three or four) produces better results. For large data sets, data balancing and the degree of dispersion have no significant impact on quality of classification. Rather, a greater number of neurons in the hidden layer produces better results (ranging from three to five times the number of neurons in the input layer).
format Online
Article
Text
id pubmed-10217427
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-102174272023-05-27 Study on the Use of Artificially Generated Objects in the Process of Training MLP Neural Networks Based on Dispersed Data Marfo, Kwabena Frimpong Przybyła-Kasperek, Małgorzata Entropy (Basel) Article This study concerns dispersed data stored in independent local tables with different sets of attributes. The paper proposes a new method for training a single neural network—a multilayer perceptron based on dispersed data. The idea is to train local models that have identical structures based on local tables; however, due to different sets of conditional attributes present in local tables, it is necessary to generate some artificial objects to train local models. The paper presents a study on the use of varying parameter values in the proposed method of creating artificial objects to train local models. The paper presents an exhaustive comparison in terms of the number of artificial objects generated based on a single original object, the degree of data dispersion, data balancing, and different network structures—the number of neurons in the hidden layer. It was found that for data sets with a large number of objects, a smaller number of artificial objects is optimal. For smaller data sets, a greater number of artificial objects (three or four) produces better results. For large data sets, data balancing and the degree of dispersion have no significant impact on quality of classification. Rather, a greater number of neurons in the hidden layer produces better results (ranging from three to five times the number of neurons in the input layer). MDPI 2023-04-24 /pmc/articles/PMC10217427/ /pubmed/37238459 http://dx.doi.org/10.3390/e25050703 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Marfo, Kwabena Frimpong
Przybyła-Kasperek, Małgorzata
Study on the Use of Artificially Generated Objects in the Process of Training MLP Neural Networks Based on Dispersed Data
title Study on the Use of Artificially Generated Objects in the Process of Training MLP Neural Networks Based on Dispersed Data
title_full Study on the Use of Artificially Generated Objects in the Process of Training MLP Neural Networks Based on Dispersed Data
title_fullStr Study on the Use of Artificially Generated Objects in the Process of Training MLP Neural Networks Based on Dispersed Data
title_full_unstemmed Study on the Use of Artificially Generated Objects in the Process of Training MLP Neural Networks Based on Dispersed Data
title_short Study on the Use of Artificially Generated Objects in the Process of Training MLP Neural Networks Based on Dispersed Data
title_sort study on the use of artificially generated objects in the process of training mlp neural networks based on dispersed data
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10217427/
https://www.ncbi.nlm.nih.gov/pubmed/37238459
http://dx.doi.org/10.3390/e25050703
work_keys_str_mv AT marfokwabenafrimpong studyontheuseofartificiallygeneratedobjectsintheprocessoftrainingmlpneuralnetworksbasedondisperseddata
AT przybyłakasperekmałgorzata studyontheuseofartificiallygeneratedobjectsintheprocessoftrainingmlpneuralnetworksbasedondisperseddata