Cargando…
Comparing methods for statistical inference with model uncertainty
Probability models are used for many statistical tasks, notably parameter estimation, interval estimation, inference about model parameters, point prediction, and interval prediction. Thus, choosing a statistical model and accounting for uncertainty about this choice are important parts of the scien...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
National Academy of Sciences
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9169744/ https://www.ncbi.nlm.nih.gov/pubmed/35412893 http://dx.doi.org/10.1073/pnas.2120737119 |
_version_ | 1784721263534538752 |
---|---|
author | Porwal, Anupreet Raftery, Adrian E. |
author_facet | Porwal, Anupreet Raftery, Adrian E. |
author_sort | Porwal, Anupreet |
collection | PubMed |
description | Probability models are used for many statistical tasks, notably parameter estimation, interval estimation, inference about model parameters, point prediction, and interval prediction. Thus, choosing a statistical model and accounting for uncertainty about this choice are important parts of the scientific process. Here we focus on one such choice, that of variables to include in a linear regression model. Many methods have been proposed, including Bayesian and penalized likelihood methods, and it is unclear which one to use. We compared 21 of the most popular methods by carrying out an extensive set of simulation studies based closely on real datasets that span a range of situations encountered in practical data analysis. Three adaptive Bayesian model averaging (BMA) methods performed best across all statistical tasks. These used adaptive versions of Zellner’s g-prior for the parameters, where the prior variance parameter g is a function of sample size or is estimated from the data. We found that for BMA methods implemented with Markov chain Monte Carlo, 10,000 iterations were enough. Computationally, we found two of the three best methods (BMA with [Formula: see text] and empirical Bayes-local) to be competitive with the least absolute shrinkage and selection operator (LASSO), which is often preferred as a variable selection technique because of its computational efficiency. BMA performed better than Bayesian model selection (in which just one model is selected). |
format | Online Article Text |
id | pubmed-9169744 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | National Academy of Sciences |
record_format | MEDLINE/PubMed |
spelling | pubmed-91697442022-06-07 Comparing methods for statistical inference with model uncertainty Porwal, Anupreet Raftery, Adrian E. Proc Natl Acad Sci U S A Physical Sciences Probability models are used for many statistical tasks, notably parameter estimation, interval estimation, inference about model parameters, point prediction, and interval prediction. Thus, choosing a statistical model and accounting for uncertainty about this choice are important parts of the scientific process. Here we focus on one such choice, that of variables to include in a linear regression model. Many methods have been proposed, including Bayesian and penalized likelihood methods, and it is unclear which one to use. We compared 21 of the most popular methods by carrying out an extensive set of simulation studies based closely on real datasets that span a range of situations encountered in practical data analysis. Three adaptive Bayesian model averaging (BMA) methods performed best across all statistical tasks. These used adaptive versions of Zellner’s g-prior for the parameters, where the prior variance parameter g is a function of sample size or is estimated from the data. We found that for BMA methods implemented with Markov chain Monte Carlo, 10,000 iterations were enough. Computationally, we found two of the three best methods (BMA with [Formula: see text] and empirical Bayes-local) to be competitive with the least absolute shrinkage and selection operator (LASSO), which is often preferred as a variable selection technique because of its computational efficiency. BMA performed better than Bayesian model selection (in which just one model is selected). National Academy of Sciences 2022-04-11 2022-04-19 /pmc/articles/PMC9169744/ /pubmed/35412893 http://dx.doi.org/10.1073/pnas.2120737119 Text en Copyright © 2022 the Author(s). Published by PNAS. https://creativecommons.org/licenses/by/4.0/This open access article is distributed under Creative Commons Attribution License 4.0 (CC BY) (https://creativecommons.org/licenses/by/4.0/) . |
spellingShingle | Physical Sciences Porwal, Anupreet Raftery, Adrian E. Comparing methods for statistical inference with model uncertainty |
title | Comparing methods for statistical inference with model uncertainty |
title_full | Comparing methods for statistical inference with model uncertainty |
title_fullStr | Comparing methods for statistical inference with model uncertainty |
title_full_unstemmed | Comparing methods for statistical inference with model uncertainty |
title_short | Comparing methods for statistical inference with model uncertainty |
title_sort | comparing methods for statistical inference with model uncertainty |
topic | Physical Sciences |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9169744/ https://www.ncbi.nlm.nih.gov/pubmed/35412893 http://dx.doi.org/10.1073/pnas.2120737119 |
work_keys_str_mv | AT porwalanupreet comparingmethodsforstatisticalinferencewithmodeluncertainty AT rafteryadriane comparingmethodsforstatisticalinferencewithmodeluncertainty |