Cargando…

Subgroup Preference Neural Network

Subgroup label ranking aims to rank groups of labels using a single ranking model, is a new problem faced in preference learning. This paper introduces the Subgroup Preference Neural Network (SGPNN) that combines multiple networks have different activation function, learning rate, and output layer i...

Descripción completa

Detalles Bibliográficos
Autores principales: Elgharabawy, Ayman, Prasad, Mukesh, Lin, Chin-Teng
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8471160/
https://www.ncbi.nlm.nih.gov/pubmed/34577312
http://dx.doi.org/10.3390/s21186104
_version_ 1784574391276797952
author Elgharabawy, Ayman
Prasad, Mukesh
Lin, Chin-Teng
author_facet Elgharabawy, Ayman
Prasad, Mukesh
Lin, Chin-Teng
author_sort Elgharabawy, Ayman
collection PubMed
description Subgroup label ranking aims to rank groups of labels using a single ranking model, is a new problem faced in preference learning. This paper introduces the Subgroup Preference Neural Network (SGPNN) that combines multiple networks have different activation function, learning rate, and output layer into one artificial neural network (ANN) to discover the hidden relation between the subgroups’ multi-labels. The SGPNN is a feedforward (FF), partially connected network that has a single middle layer and uses stairstep (SS) multi-valued activation function to enhance the prediction’s probability and accelerate the ranking convergence. The novel structure of the proposed SGPNN consists of a multi-activation function neuron (MAFN) in the middle layer to rank each subgroup independently. The SGPNN uses gradient ascent to maximize the Spearman ranking correlation between the groups of labels. Each label is represented by an output neuron that has a single SS function. The proposed SGPNN using conjoint dataset outperforms the other label ranking methods which uses each dataset individually. The proposed SGPNN achieves an average accuracy of 91.4% using the conjoint dataset compared to supervised clustering, decision tree, multilayer perceptron label ranking and label ranking forests that achieve an average accuracy of 60%, 84.8%, 69.2% and 73%, respectively, using the individual dataset.
format Online
Article
Text
id pubmed-8471160
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-84711602021-09-27 Subgroup Preference Neural Network Elgharabawy, Ayman Prasad, Mukesh Lin, Chin-Teng Sensors (Basel) Communication Subgroup label ranking aims to rank groups of labels using a single ranking model, is a new problem faced in preference learning. This paper introduces the Subgroup Preference Neural Network (SGPNN) that combines multiple networks have different activation function, learning rate, and output layer into one artificial neural network (ANN) to discover the hidden relation between the subgroups’ multi-labels. The SGPNN is a feedforward (FF), partially connected network that has a single middle layer and uses stairstep (SS) multi-valued activation function to enhance the prediction’s probability and accelerate the ranking convergence. The novel structure of the proposed SGPNN consists of a multi-activation function neuron (MAFN) in the middle layer to rank each subgroup independently. The SGPNN uses gradient ascent to maximize the Spearman ranking correlation between the groups of labels. Each label is represented by an output neuron that has a single SS function. The proposed SGPNN using conjoint dataset outperforms the other label ranking methods which uses each dataset individually. The proposed SGPNN achieves an average accuracy of 91.4% using the conjoint dataset compared to supervised clustering, decision tree, multilayer perceptron label ranking and label ranking forests that achieve an average accuracy of 60%, 84.8%, 69.2% and 73%, respectively, using the individual dataset. MDPI 2021-09-12 /pmc/articles/PMC8471160/ /pubmed/34577312 http://dx.doi.org/10.3390/s21186104 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Communication
Elgharabawy, Ayman
Prasad, Mukesh
Lin, Chin-Teng
Subgroup Preference Neural Network
title Subgroup Preference Neural Network
title_full Subgroup Preference Neural Network
title_fullStr Subgroup Preference Neural Network
title_full_unstemmed Subgroup Preference Neural Network
title_short Subgroup Preference Neural Network
title_sort subgroup preference neural network
topic Communication
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8471160/
https://www.ncbi.nlm.nih.gov/pubmed/34577312
http://dx.doi.org/10.3390/s21186104
work_keys_str_mv AT elgharabawyayman subgrouppreferenceneuralnetwork
AT prasadmukesh subgrouppreferenceneuralnetwork
AT linchinteng subgrouppreferenceneuralnetwork