Cargando…

Posterior Averaging Information Criterion

We propose a new model selection method, named the posterior averaging information criterion, for Bayesian model assessment to minimize the risk of predicting independent future observations. The theoretical foundation is built on the Kullback–Leibler divergence to quantify the similarity between th...

Descripción completa

Detalles Bibliográficos
Autor principal: Zhou, Shouhao
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10047922/
https://www.ncbi.nlm.nih.gov/pubmed/36981356
http://dx.doi.org/10.3390/e25030468
_version_ 1785014048269533184
author Zhou, Shouhao
author_facet Zhou, Shouhao
author_sort Zhou, Shouhao
collection PubMed
description We propose a new model selection method, named the posterior averaging information criterion, for Bayesian model assessment to minimize the risk of predicting independent future observations. The theoretical foundation is built on the Kullback–Leibler divergence to quantify the similarity between the proposed candidate model and the underlying true model. From a Bayesian perspective, our method evaluates the candidate models over the entire posterior distribution in terms of predicting a future independent observation. Without assuming that the true distribution is contained in the candidate models, the new criterion is developed by correcting the asymptotic bias of the posterior mean of the in-sample log-likelihood against out-of-sample log-likelihood, and can be generally applied even for Bayesian models with degenerate non-informative priors. Simulations in both normal and binomial settings demonstrate superior small sample performance.
format Online
Article
Text
id pubmed-10047922
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-100479222023-03-29 Posterior Averaging Information Criterion Zhou, Shouhao Entropy (Basel) Article We propose a new model selection method, named the posterior averaging information criterion, for Bayesian model assessment to minimize the risk of predicting independent future observations. The theoretical foundation is built on the Kullback–Leibler divergence to quantify the similarity between the proposed candidate model and the underlying true model. From a Bayesian perspective, our method evaluates the candidate models over the entire posterior distribution in terms of predicting a future independent observation. Without assuming that the true distribution is contained in the candidate models, the new criterion is developed by correcting the asymptotic bias of the posterior mean of the in-sample log-likelihood against out-of-sample log-likelihood, and can be generally applied even for Bayesian models with degenerate non-informative priors. Simulations in both normal and binomial settings demonstrate superior small sample performance. MDPI 2023-03-07 /pmc/articles/PMC10047922/ /pubmed/36981356 http://dx.doi.org/10.3390/e25030468 Text en © 2023 by the author. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Zhou, Shouhao
Posterior Averaging Information Criterion
title Posterior Averaging Information Criterion
title_full Posterior Averaging Information Criterion
title_fullStr Posterior Averaging Information Criterion
title_full_unstemmed Posterior Averaging Information Criterion
title_short Posterior Averaging Information Criterion
title_sort posterior averaging information criterion
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10047922/
https://www.ncbi.nlm.nih.gov/pubmed/36981356
http://dx.doi.org/10.3390/e25030468
work_keys_str_mv AT zhoushouhao posterioraveraginginformationcriterion