Cargando…

Mixture-Based Probabilistic Graphical Models for the Label Ranking Problem †

The goal of the Label Ranking (LR) problem is to learn preference models that predict the preferred ranking of class labels for a given unlabeled instance. Different well-known machine learning algorithms have been adapted to deal with the LR problem. In particular, fine-tuned instance-based algorit...

Descripción completa

Detalles Bibliográficos
Autores principales: Rodrigo, Enrique G., Alfaro, Juan C., Aledo, Juan A., Gámez, José A.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8066248/
https://www.ncbi.nlm.nih.gov/pubmed/33807440
http://dx.doi.org/10.3390/e23040420
_version_ 1783682530203926528
author Rodrigo, Enrique G.
Alfaro, Juan C.
Aledo, Juan A.
Gámez, José A.
author_facet Rodrigo, Enrique G.
Alfaro, Juan C.
Aledo, Juan A.
Gámez, José A.
author_sort Rodrigo, Enrique G.
collection PubMed
description The goal of the Label Ranking (LR) problem is to learn preference models that predict the preferred ranking of class labels for a given unlabeled instance. Different well-known machine learning algorithms have been adapted to deal with the LR problem. In particular, fine-tuned instance-based algorithms (e.g., k-nearest neighbors) and model-based algorithms (e.g., decision trees) have performed remarkably well in tackling the LR problem. Probabilistic Graphical Models (PGMs, e.g., Bayesian networks) have not been considered to deal with this problem because of the difficulty of modeling permutations in that framework. In this paper, we propose a Hidden Naive Bayes classifier (HNB) to cope with the LR problem. By introducing a hidden variable, we can design a hybrid Bayesian network in which several types of distributions can be combined: multinomial for discrete variables, Gaussian for numerical variables, and Mallows for permutations. We consider two kinds of probabilistic models: one based on a Naive Bayes graphical structure (where only univariate probability distributions are estimated for each state of the hidden variable) and another where we allow interactions among the predictive attributes (using a multivariate Gaussian distribution for the parameter estimation). The experimental evaluation shows that our proposals are competitive with the start-of-the-art algorithms in both accuracy and in CPU time requirements.
format Online
Article
Text
id pubmed-8066248
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-80662482021-04-25 Mixture-Based Probabilistic Graphical Models for the Label Ranking Problem † Rodrigo, Enrique G. Alfaro, Juan C. Aledo, Juan A. Gámez, José A. Entropy (Basel) Article The goal of the Label Ranking (LR) problem is to learn preference models that predict the preferred ranking of class labels for a given unlabeled instance. Different well-known machine learning algorithms have been adapted to deal with the LR problem. In particular, fine-tuned instance-based algorithms (e.g., k-nearest neighbors) and model-based algorithms (e.g., decision trees) have performed remarkably well in tackling the LR problem. Probabilistic Graphical Models (PGMs, e.g., Bayesian networks) have not been considered to deal with this problem because of the difficulty of modeling permutations in that framework. In this paper, we propose a Hidden Naive Bayes classifier (HNB) to cope with the LR problem. By introducing a hidden variable, we can design a hybrid Bayesian network in which several types of distributions can be combined: multinomial for discrete variables, Gaussian for numerical variables, and Mallows for permutations. We consider two kinds of probabilistic models: one based on a Naive Bayes graphical structure (where only univariate probability distributions are estimated for each state of the hidden variable) and another where we allow interactions among the predictive attributes (using a multivariate Gaussian distribution for the parameter estimation). The experimental evaluation shows that our proposals are competitive with the start-of-the-art algorithms in both accuracy and in CPU time requirements. MDPI 2021-03-31 /pmc/articles/PMC8066248/ /pubmed/33807440 http://dx.doi.org/10.3390/e23040420 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Rodrigo, Enrique G.
Alfaro, Juan C.
Aledo, Juan A.
Gámez, José A.
Mixture-Based Probabilistic Graphical Models for the Label Ranking Problem †
title Mixture-Based Probabilistic Graphical Models for the Label Ranking Problem †
title_full Mixture-Based Probabilistic Graphical Models for the Label Ranking Problem †
title_fullStr Mixture-Based Probabilistic Graphical Models for the Label Ranking Problem †
title_full_unstemmed Mixture-Based Probabilistic Graphical Models for the Label Ranking Problem †
title_short Mixture-Based Probabilistic Graphical Models for the Label Ranking Problem †
title_sort mixture-based probabilistic graphical models for the label ranking problem †
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8066248/
https://www.ncbi.nlm.nih.gov/pubmed/33807440
http://dx.doi.org/10.3390/e23040420
work_keys_str_mv AT rodrigoenriqueg mixturebasedprobabilisticgraphicalmodelsforthelabelrankingproblem
AT alfarojuanc mixturebasedprobabilisticgraphicalmodelsforthelabelrankingproblem
AT aledojuana mixturebasedprobabilisticgraphicalmodelsforthelabelrankingproblem
AT gamezjosea mixturebasedprobabilisticgraphicalmodelsforthelabelrankingproblem