Cargando…

Fighting or embracing multiplicity in neuroimaging? neighborhood leverage versus global calibration

Neuroimaging faces the daunting challenge of multiple testing – an instance of multiplicity – that is associated with two other issues to some extent: low inference efficiency and poor reproducibility. Typically, the same statistical model is applied to each spatial unit independently in the approac...

Descripción completa

Detalles Bibliográficos
Autores principales: Chen, Gang, Taylor, Paul A., Cox, Robert W., Pessoa, Luiz
Formato: Online Artículo Texto
Lenguaje:English
Publicado: 2019
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6980934/
https://www.ncbi.nlm.nih.gov/pubmed/31698079
http://dx.doi.org/10.1016/j.neuroimage.2019.116320
_version_ 1783491009393459200
author Chen, Gang
Taylor, Paul A.
Cox, Robert W.
Pessoa, Luiz
author_facet Chen, Gang
Taylor, Paul A.
Cox, Robert W.
Pessoa, Luiz
author_sort Chen, Gang
collection PubMed
description Neuroimaging faces the daunting challenge of multiple testing – an instance of multiplicity – that is associated with two other issues to some extent: low inference efficiency and poor reproducibility. Typically, the same statistical model is applied to each spatial unit independently in the approach of massively univariate modeling. In dealing with multiplicity, the general strategy employed in the field is the same regardless of the specifics: trust the local “unbiased” effect estimates while adjusting the extent of statistical evidence at the global level. However, in this approach, modeling efficiency is compromised because each spatial unit (e.g., voxel, region, matrix element) is treated as an isolated and independent entity during massively univariate modeling. In addition, the required step of multiple testing “correction” by taking into consideration spatial relatedness, or neighborhood leverage, can only partly recoup statistical efficiency, resulting in potentially excessive penalization as well as arbitrariness due to thresholding procedures. Moreover, the assigned statistical evidence at the global level heavily relies on the data space (whole brain or a small volume). The present paper reviews how Stein’s paradox (1956) motivates a Bayesian multilevel (BML) approach that, rather than fighting multiplicity, embraces it to our advantage through a global calibration process among spatial units. Global calibration is accomplished via a Gaussian distribution for the cross-region effects whose properties are not a priori specified, but a posteriori determined by the data at hand through the BML model. Our framework therefore incorporates multiplicity as integral to the modeling structure, not a separate correction step. By turning multiplicity into a strength, we aim to achieve five goals: 1) improve the model efficiency with a higher predictive accuracy, 2) control the errors of incorrect magnitude and incorrect sign, 3) validate each model relative to competing candidates, 4) reduce the reliance and sensitivity on the choice of data space, and 5) encourage full results reporting. Our modeling proposal reverberates with recent proposals to eliminate the dichotomization of statistical evidence (“significant” vs. “non-significant”), to improve the interpretability of study findings, as well as to promote reporting the full gamut of results (not only “significant” ones), thereby enhancing research transparency and reproducibility.
format Online
Article
Text
id pubmed-6980934
institution National Center for Biotechnology Information
language English
publishDate 2019
record_format MEDLINE/PubMed
spelling pubmed-69809342021-02-01 Fighting or embracing multiplicity in neuroimaging? neighborhood leverage versus global calibration Chen, Gang Taylor, Paul A. Cox, Robert W. Pessoa, Luiz Neuroimage Article Neuroimaging faces the daunting challenge of multiple testing – an instance of multiplicity – that is associated with two other issues to some extent: low inference efficiency and poor reproducibility. Typically, the same statistical model is applied to each spatial unit independently in the approach of massively univariate modeling. In dealing with multiplicity, the general strategy employed in the field is the same regardless of the specifics: trust the local “unbiased” effect estimates while adjusting the extent of statistical evidence at the global level. However, in this approach, modeling efficiency is compromised because each spatial unit (e.g., voxel, region, matrix element) is treated as an isolated and independent entity during massively univariate modeling. In addition, the required step of multiple testing “correction” by taking into consideration spatial relatedness, or neighborhood leverage, can only partly recoup statistical efficiency, resulting in potentially excessive penalization as well as arbitrariness due to thresholding procedures. Moreover, the assigned statistical evidence at the global level heavily relies on the data space (whole brain or a small volume). The present paper reviews how Stein’s paradox (1956) motivates a Bayesian multilevel (BML) approach that, rather than fighting multiplicity, embraces it to our advantage through a global calibration process among spatial units. Global calibration is accomplished via a Gaussian distribution for the cross-region effects whose properties are not a priori specified, but a posteriori determined by the data at hand through the BML model. Our framework therefore incorporates multiplicity as integral to the modeling structure, not a separate correction step. By turning multiplicity into a strength, we aim to achieve five goals: 1) improve the model efficiency with a higher predictive accuracy, 2) control the errors of incorrect magnitude and incorrect sign, 3) validate each model relative to competing candidates, 4) reduce the reliance and sensitivity on the choice of data space, and 5) encourage full results reporting. Our modeling proposal reverberates with recent proposals to eliminate the dichotomization of statistical evidence (“significant” vs. “non-significant”), to improve the interpretability of study findings, as well as to promote reporting the full gamut of results (not only “significant” ones), thereby enhancing research transparency and reproducibility. 2019-11-05 2020-02-01 /pmc/articles/PMC6980934/ /pubmed/31698079 http://dx.doi.org/10.1016/j.neuroimage.2019.116320 Text en This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/).
spellingShingle Article
Chen, Gang
Taylor, Paul A.
Cox, Robert W.
Pessoa, Luiz
Fighting or embracing multiplicity in neuroimaging? neighborhood leverage versus global calibration
title Fighting or embracing multiplicity in neuroimaging? neighborhood leverage versus global calibration
title_full Fighting or embracing multiplicity in neuroimaging? neighborhood leverage versus global calibration
title_fullStr Fighting or embracing multiplicity in neuroimaging? neighborhood leverage versus global calibration
title_full_unstemmed Fighting or embracing multiplicity in neuroimaging? neighborhood leverage versus global calibration
title_short Fighting or embracing multiplicity in neuroimaging? neighborhood leverage versus global calibration
title_sort fighting or embracing multiplicity in neuroimaging? neighborhood leverage versus global calibration
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6980934/
https://www.ncbi.nlm.nih.gov/pubmed/31698079
http://dx.doi.org/10.1016/j.neuroimage.2019.116320
work_keys_str_mv AT chengang fightingorembracingmultiplicityinneuroimagingneighborhoodleverageversusglobalcalibration
AT taylorpaula fightingorembracingmultiplicityinneuroimagingneighborhoodleverageversusglobalcalibration
AT coxrobertw fightingorembracingmultiplicityinneuroimagingneighborhoodleverageversusglobalcalibration
AT pessoaluiz fightingorembracingmultiplicityinneuroimagingneighborhoodleverageversusglobalcalibration