Cargando…
Fighting or embracing multiplicity in neuroimaging? neighborhood leverage versus global calibration
Neuroimaging faces the daunting challenge of multiple testing – an instance of multiplicity – that is associated with two other issues to some extent: low inference efficiency and poor reproducibility. Typically, the same statistical model is applied to each spatial unit independently in the approac...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
2019
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6980934/ https://www.ncbi.nlm.nih.gov/pubmed/31698079 http://dx.doi.org/10.1016/j.neuroimage.2019.116320 |
_version_ | 1783491009393459200 |
---|---|
author | Chen, Gang Taylor, Paul A. Cox, Robert W. Pessoa, Luiz |
author_facet | Chen, Gang Taylor, Paul A. Cox, Robert W. Pessoa, Luiz |
author_sort | Chen, Gang |
collection | PubMed |
description | Neuroimaging faces the daunting challenge of multiple testing – an instance of multiplicity – that is associated with two other issues to some extent: low inference efficiency and poor reproducibility. Typically, the same statistical model is applied to each spatial unit independently in the approach of massively univariate modeling. In dealing with multiplicity, the general strategy employed in the field is the same regardless of the specifics: trust the local “unbiased” effect estimates while adjusting the extent of statistical evidence at the global level. However, in this approach, modeling efficiency is compromised because each spatial unit (e.g., voxel, region, matrix element) is treated as an isolated and independent entity during massively univariate modeling. In addition, the required step of multiple testing “correction” by taking into consideration spatial relatedness, or neighborhood leverage, can only partly recoup statistical efficiency, resulting in potentially excessive penalization as well as arbitrariness due to thresholding procedures. Moreover, the assigned statistical evidence at the global level heavily relies on the data space (whole brain or a small volume). The present paper reviews how Stein’s paradox (1956) motivates a Bayesian multilevel (BML) approach that, rather than fighting multiplicity, embraces it to our advantage through a global calibration process among spatial units. Global calibration is accomplished via a Gaussian distribution for the cross-region effects whose properties are not a priori specified, but a posteriori determined by the data at hand through the BML model. Our framework therefore incorporates multiplicity as integral to the modeling structure, not a separate correction step. By turning multiplicity into a strength, we aim to achieve five goals: 1) improve the model efficiency with a higher predictive accuracy, 2) control the errors of incorrect magnitude and incorrect sign, 3) validate each model relative to competing candidates, 4) reduce the reliance and sensitivity on the choice of data space, and 5) encourage full results reporting. Our modeling proposal reverberates with recent proposals to eliminate the dichotomization of statistical evidence (“significant” vs. “non-significant”), to improve the interpretability of study findings, as well as to promote reporting the full gamut of results (not only “significant” ones), thereby enhancing research transparency and reproducibility. |
format | Online Article Text |
id | pubmed-6980934 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2019 |
record_format | MEDLINE/PubMed |
spelling | pubmed-69809342021-02-01 Fighting or embracing multiplicity in neuroimaging? neighborhood leverage versus global calibration Chen, Gang Taylor, Paul A. Cox, Robert W. Pessoa, Luiz Neuroimage Article Neuroimaging faces the daunting challenge of multiple testing – an instance of multiplicity – that is associated with two other issues to some extent: low inference efficiency and poor reproducibility. Typically, the same statistical model is applied to each spatial unit independently in the approach of massively univariate modeling. In dealing with multiplicity, the general strategy employed in the field is the same regardless of the specifics: trust the local “unbiased” effect estimates while adjusting the extent of statistical evidence at the global level. However, in this approach, modeling efficiency is compromised because each spatial unit (e.g., voxel, region, matrix element) is treated as an isolated and independent entity during massively univariate modeling. In addition, the required step of multiple testing “correction” by taking into consideration spatial relatedness, or neighborhood leverage, can only partly recoup statistical efficiency, resulting in potentially excessive penalization as well as arbitrariness due to thresholding procedures. Moreover, the assigned statistical evidence at the global level heavily relies on the data space (whole brain or a small volume). The present paper reviews how Stein’s paradox (1956) motivates a Bayesian multilevel (BML) approach that, rather than fighting multiplicity, embraces it to our advantage through a global calibration process among spatial units. Global calibration is accomplished via a Gaussian distribution for the cross-region effects whose properties are not a priori specified, but a posteriori determined by the data at hand through the BML model. Our framework therefore incorporates multiplicity as integral to the modeling structure, not a separate correction step. By turning multiplicity into a strength, we aim to achieve five goals: 1) improve the model efficiency with a higher predictive accuracy, 2) control the errors of incorrect magnitude and incorrect sign, 3) validate each model relative to competing candidates, 4) reduce the reliance and sensitivity on the choice of data space, and 5) encourage full results reporting. Our modeling proposal reverberates with recent proposals to eliminate the dichotomization of statistical evidence (“significant” vs. “non-significant”), to improve the interpretability of study findings, as well as to promote reporting the full gamut of results (not only “significant” ones), thereby enhancing research transparency and reproducibility. 2019-11-05 2020-02-01 /pmc/articles/PMC6980934/ /pubmed/31698079 http://dx.doi.org/10.1016/j.neuroimage.2019.116320 Text en This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/). |
spellingShingle | Article Chen, Gang Taylor, Paul A. Cox, Robert W. Pessoa, Luiz Fighting or embracing multiplicity in neuroimaging? neighborhood leverage versus global calibration |
title | Fighting or embracing multiplicity in neuroimaging? neighborhood leverage versus global calibration |
title_full | Fighting or embracing multiplicity in neuroimaging? neighborhood leverage versus global calibration |
title_fullStr | Fighting or embracing multiplicity in neuroimaging? neighborhood leverage versus global calibration |
title_full_unstemmed | Fighting or embracing multiplicity in neuroimaging? neighborhood leverage versus global calibration |
title_short | Fighting or embracing multiplicity in neuroimaging? neighborhood leverage versus global calibration |
title_sort | fighting or embracing multiplicity in neuroimaging? neighborhood leverage versus global calibration |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6980934/ https://www.ncbi.nlm.nih.gov/pubmed/31698079 http://dx.doi.org/10.1016/j.neuroimage.2019.116320 |
work_keys_str_mv | AT chengang fightingorembracingmultiplicityinneuroimagingneighborhoodleverageversusglobalcalibration AT taylorpaula fightingorembracingmultiplicityinneuroimagingneighborhoodleverageversusglobalcalibration AT coxrobertw fightingorembracingmultiplicityinneuroimagingneighborhoodleverageversusglobalcalibration AT pessoaluiz fightingorembracingmultiplicityinneuroimagingneighborhoodleverageversusglobalcalibration |