Cargando…

How Good Is Crude MDL for Solving the Bias-Variance Dilemma? An Empirical Investigation Based on Bayesian Networks

The bias-variance dilemma is a well-known and important problem in Machine Learning. It basically relates the generalization capability (goodness of fit) of a learning method to its corresponding complexity. When we have enough data at hand, it is possible to use these data in such a way so as to mi...

Descripción completa

Detalles Bibliográficos
Autores principales: Cruz-Ramírez, Nicandro, Acosta-Mesa, Héctor Gabriel, Mezura-Montes, Efrén, Guerra-Hernández, Alejandro, Hoyos-Rivera, Guillermo de Jesús, Barrientos-Martínez, Rocío Erandi, Gutiérrez-Fragoso, Karina, Nava-Fernández, Luis Alonso, González-Gaspar, Patricia, Novoa-del-Toro, Elva María, Aguilera-Rueda, Vicente Josué, Ameca-Alducin, María Yaneli
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Public Library of Science 2014
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3966834/
https://www.ncbi.nlm.nih.gov/pubmed/24671204
http://dx.doi.org/10.1371/journal.pone.0092866
_version_ 1782308951496327168
author Cruz-Ramírez, Nicandro
Acosta-Mesa, Héctor Gabriel
Mezura-Montes, Efrén
Guerra-Hernández, Alejandro
Hoyos-Rivera, Guillermo de Jesús
Barrientos-Martínez, Rocío Erandi
Gutiérrez-Fragoso, Karina
Nava-Fernández, Luis Alonso
González-Gaspar, Patricia
Novoa-del-Toro, Elva María
Aguilera-Rueda, Vicente Josué
Ameca-Alducin, María Yaneli
author_facet Cruz-Ramírez, Nicandro
Acosta-Mesa, Héctor Gabriel
Mezura-Montes, Efrén
Guerra-Hernández, Alejandro
Hoyos-Rivera, Guillermo de Jesús
Barrientos-Martínez, Rocío Erandi
Gutiérrez-Fragoso, Karina
Nava-Fernández, Luis Alonso
González-Gaspar, Patricia
Novoa-del-Toro, Elva María
Aguilera-Rueda, Vicente Josué
Ameca-Alducin, María Yaneli
author_sort Cruz-Ramírez, Nicandro
collection PubMed
description The bias-variance dilemma is a well-known and important problem in Machine Learning. It basically relates the generalization capability (goodness of fit) of a learning method to its corresponding complexity. When we have enough data at hand, it is possible to use these data in such a way so as to minimize overfitting (the risk of selecting a complex model that generalizes poorly). Unfortunately, there are many situations where we simply do not have this required amount of data. Thus, we need to find methods capable of efficiently exploiting the available data while avoiding overfitting. Different metrics have been proposed to achieve this goal: the Minimum Description Length principle (MDL), Akaike’s Information Criterion (AIC) and Bayesian Information Criterion (BIC), among others. In this paper, we focus on crude MDL and empirically evaluate its performance in selecting models with a good balance between goodness of fit and complexity: the so-called bias-variance dilemma, decomposition or tradeoff. Although the graphical interaction between these dimensions (bias and variance) is ubiquitous in the Machine Learning literature, few works present experimental evidence to recover such interaction. In our experiments, we argue that the resulting graphs allow us to gain insights that are difficult to unveil otherwise: that crude MDL naturally selects balanced models in terms of bias-variance, which not necessarily need be the gold-standard ones. We carry out these experiments using a specific model: a Bayesian network. In spite of these motivating results, we also should not overlook three other components that may significantly affect the final model selection: the search procedure, the noise rate and the sample size.
format Online
Article
Text
id pubmed-3966834
institution National Center for Biotechnology Information
language English
publishDate 2014
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-39668342014-03-31 How Good Is Crude MDL for Solving the Bias-Variance Dilemma? An Empirical Investigation Based on Bayesian Networks Cruz-Ramírez, Nicandro Acosta-Mesa, Héctor Gabriel Mezura-Montes, Efrén Guerra-Hernández, Alejandro Hoyos-Rivera, Guillermo de Jesús Barrientos-Martínez, Rocío Erandi Gutiérrez-Fragoso, Karina Nava-Fernández, Luis Alonso González-Gaspar, Patricia Novoa-del-Toro, Elva María Aguilera-Rueda, Vicente Josué Ameca-Alducin, María Yaneli PLoS One Research Article The bias-variance dilemma is a well-known and important problem in Machine Learning. It basically relates the generalization capability (goodness of fit) of a learning method to its corresponding complexity. When we have enough data at hand, it is possible to use these data in such a way so as to minimize overfitting (the risk of selecting a complex model that generalizes poorly). Unfortunately, there are many situations where we simply do not have this required amount of data. Thus, we need to find methods capable of efficiently exploiting the available data while avoiding overfitting. Different metrics have been proposed to achieve this goal: the Minimum Description Length principle (MDL), Akaike’s Information Criterion (AIC) and Bayesian Information Criterion (BIC), among others. In this paper, we focus on crude MDL and empirically evaluate its performance in selecting models with a good balance between goodness of fit and complexity: the so-called bias-variance dilemma, decomposition or tradeoff. Although the graphical interaction between these dimensions (bias and variance) is ubiquitous in the Machine Learning literature, few works present experimental evidence to recover such interaction. In our experiments, we argue that the resulting graphs allow us to gain insights that are difficult to unveil otherwise: that crude MDL naturally selects balanced models in terms of bias-variance, which not necessarily need be the gold-standard ones. We carry out these experiments using a specific model: a Bayesian network. In spite of these motivating results, we also should not overlook three other components that may significantly affect the final model selection: the search procedure, the noise rate and the sample size. Public Library of Science 2014-03-26 /pmc/articles/PMC3966834/ /pubmed/24671204 http://dx.doi.org/10.1371/journal.pone.0092866 Text en © 2014 Cruz-Ramírez et al http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are properly credited.
spellingShingle Research Article
Cruz-Ramírez, Nicandro
Acosta-Mesa, Héctor Gabriel
Mezura-Montes, Efrén
Guerra-Hernández, Alejandro
Hoyos-Rivera, Guillermo de Jesús
Barrientos-Martínez, Rocío Erandi
Gutiérrez-Fragoso, Karina
Nava-Fernández, Luis Alonso
González-Gaspar, Patricia
Novoa-del-Toro, Elva María
Aguilera-Rueda, Vicente Josué
Ameca-Alducin, María Yaneli
How Good Is Crude MDL for Solving the Bias-Variance Dilemma? An Empirical Investigation Based on Bayesian Networks
title How Good Is Crude MDL for Solving the Bias-Variance Dilemma? An Empirical Investigation Based on Bayesian Networks
title_full How Good Is Crude MDL for Solving the Bias-Variance Dilemma? An Empirical Investigation Based on Bayesian Networks
title_fullStr How Good Is Crude MDL for Solving the Bias-Variance Dilemma? An Empirical Investigation Based on Bayesian Networks
title_full_unstemmed How Good Is Crude MDL for Solving the Bias-Variance Dilemma? An Empirical Investigation Based on Bayesian Networks
title_short How Good Is Crude MDL for Solving the Bias-Variance Dilemma? An Empirical Investigation Based on Bayesian Networks
title_sort how good is crude mdl for solving the bias-variance dilemma? an empirical investigation based on bayesian networks
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3966834/
https://www.ncbi.nlm.nih.gov/pubmed/24671204
http://dx.doi.org/10.1371/journal.pone.0092866
work_keys_str_mv AT cruzramireznicandro howgoodiscrudemdlforsolvingthebiasvariancedilemmaanempiricalinvestigationbasedonbayesiannetworks
AT acostamesahectorgabriel howgoodiscrudemdlforsolvingthebiasvariancedilemmaanempiricalinvestigationbasedonbayesiannetworks
AT mezuramontesefren howgoodiscrudemdlforsolvingthebiasvariancedilemmaanempiricalinvestigationbasedonbayesiannetworks
AT guerrahernandezalejandro howgoodiscrudemdlforsolvingthebiasvariancedilemmaanempiricalinvestigationbasedonbayesiannetworks
AT hoyosriveraguillermodejesus howgoodiscrudemdlforsolvingthebiasvariancedilemmaanempiricalinvestigationbasedonbayesiannetworks
AT barrientosmartinezrocioerandi howgoodiscrudemdlforsolvingthebiasvariancedilemmaanempiricalinvestigationbasedonbayesiannetworks
AT gutierrezfragosokarina howgoodiscrudemdlforsolvingthebiasvariancedilemmaanempiricalinvestigationbasedonbayesiannetworks
AT navafernandezluisalonso howgoodiscrudemdlforsolvingthebiasvariancedilemmaanempiricalinvestigationbasedonbayesiannetworks
AT gonzalezgasparpatricia howgoodiscrudemdlforsolvingthebiasvariancedilemmaanempiricalinvestigationbasedonbayesiannetworks
AT novoadeltoroelvamaria howgoodiscrudemdlforsolvingthebiasvariancedilemmaanempiricalinvestigationbasedonbayesiannetworks
AT aguileraruedavicentejosue howgoodiscrudemdlforsolvingthebiasvariancedilemmaanempiricalinvestigationbasedonbayesiannetworks
AT amecaalducinmariayaneli howgoodiscrudemdlforsolvingthebiasvariancedilemmaanempiricalinvestigationbasedonbayesiannetworks