Cargando…
Mixture-Based Probabilistic Graphical Models for the Label Ranking Problem †
The goal of the Label Ranking (LR) problem is to learn preference models that predict the preferred ranking of class labels for a given unlabeled instance. Different well-known machine learning algorithms have been adapted to deal with the LR problem. In particular, fine-tuned instance-based algorit...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8066248/ https://www.ncbi.nlm.nih.gov/pubmed/33807440 http://dx.doi.org/10.3390/e23040420 |
_version_ | 1783682530203926528 |
---|---|
author | Rodrigo, Enrique G. Alfaro, Juan C. Aledo, Juan A. Gámez, José A. |
author_facet | Rodrigo, Enrique G. Alfaro, Juan C. Aledo, Juan A. Gámez, José A. |
author_sort | Rodrigo, Enrique G. |
collection | PubMed |
description | The goal of the Label Ranking (LR) problem is to learn preference models that predict the preferred ranking of class labels for a given unlabeled instance. Different well-known machine learning algorithms have been adapted to deal with the LR problem. In particular, fine-tuned instance-based algorithms (e.g., k-nearest neighbors) and model-based algorithms (e.g., decision trees) have performed remarkably well in tackling the LR problem. Probabilistic Graphical Models (PGMs, e.g., Bayesian networks) have not been considered to deal with this problem because of the difficulty of modeling permutations in that framework. In this paper, we propose a Hidden Naive Bayes classifier (HNB) to cope with the LR problem. By introducing a hidden variable, we can design a hybrid Bayesian network in which several types of distributions can be combined: multinomial for discrete variables, Gaussian for numerical variables, and Mallows for permutations. We consider two kinds of probabilistic models: one based on a Naive Bayes graphical structure (where only univariate probability distributions are estimated for each state of the hidden variable) and another where we allow interactions among the predictive attributes (using a multivariate Gaussian distribution for the parameter estimation). The experimental evaluation shows that our proposals are competitive with the start-of-the-art algorithms in both accuracy and in CPU time requirements. |
format | Online Article Text |
id | pubmed-8066248 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-80662482021-04-25 Mixture-Based Probabilistic Graphical Models for the Label Ranking Problem † Rodrigo, Enrique G. Alfaro, Juan C. Aledo, Juan A. Gámez, José A. Entropy (Basel) Article The goal of the Label Ranking (LR) problem is to learn preference models that predict the preferred ranking of class labels for a given unlabeled instance. Different well-known machine learning algorithms have been adapted to deal with the LR problem. In particular, fine-tuned instance-based algorithms (e.g., k-nearest neighbors) and model-based algorithms (e.g., decision trees) have performed remarkably well in tackling the LR problem. Probabilistic Graphical Models (PGMs, e.g., Bayesian networks) have not been considered to deal with this problem because of the difficulty of modeling permutations in that framework. In this paper, we propose a Hidden Naive Bayes classifier (HNB) to cope with the LR problem. By introducing a hidden variable, we can design a hybrid Bayesian network in which several types of distributions can be combined: multinomial for discrete variables, Gaussian for numerical variables, and Mallows for permutations. We consider two kinds of probabilistic models: one based on a Naive Bayes graphical structure (where only univariate probability distributions are estimated for each state of the hidden variable) and another where we allow interactions among the predictive attributes (using a multivariate Gaussian distribution for the parameter estimation). The experimental evaluation shows that our proposals are competitive with the start-of-the-art algorithms in both accuracy and in CPU time requirements. MDPI 2021-03-31 /pmc/articles/PMC8066248/ /pubmed/33807440 http://dx.doi.org/10.3390/e23040420 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Rodrigo, Enrique G. Alfaro, Juan C. Aledo, Juan A. Gámez, José A. Mixture-Based Probabilistic Graphical Models for the Label Ranking Problem † |
title | Mixture-Based Probabilistic Graphical Models for the Label Ranking Problem † |
title_full | Mixture-Based Probabilistic Graphical Models for the Label Ranking Problem † |
title_fullStr | Mixture-Based Probabilistic Graphical Models for the Label Ranking Problem † |
title_full_unstemmed | Mixture-Based Probabilistic Graphical Models for the Label Ranking Problem † |
title_short | Mixture-Based Probabilistic Graphical Models for the Label Ranking Problem † |
title_sort | mixture-based probabilistic graphical models for the label ranking problem † |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8066248/ https://www.ncbi.nlm.nih.gov/pubmed/33807440 http://dx.doi.org/10.3390/e23040420 |
work_keys_str_mv | AT rodrigoenriqueg mixturebasedprobabilisticgraphicalmodelsforthelabelrankingproblem AT alfarojuanc mixturebasedprobabilisticgraphicalmodelsforthelabelrankingproblem AT aledojuana mixturebasedprobabilisticgraphicalmodelsforthelabelrankingproblem AT gamezjosea mixturebasedprobabilisticgraphicalmodelsforthelabelrankingproblem |