Cargando…

Methods to split cognitive task data for estimating split-half reliability: A comprehensive review and systematic assessment

Estimating the reliability of cognitive task datasets is commonly done via split-half methods. We review four methods that differ in how the trials are split into parts: a first-second half split, an odd-even trial split, a permutated split, and a Monte Carlo-based split. Additionally, each splittin...

Descripción completa

Detalles Bibliográficos
Autores principales: Pronk, Thomas, Molenaar, Dylan, Wiers, Reinout W., Murre, Jaap
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Springer US 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8858277/
https://www.ncbi.nlm.nih.gov/pubmed/34100223
http://dx.doi.org/10.3758/s13423-021-01948-3
_version_ 1784654216148549632
author Pronk, Thomas
Molenaar, Dylan
Wiers, Reinout W.
Murre, Jaap
author_facet Pronk, Thomas
Molenaar, Dylan
Wiers, Reinout W.
Murre, Jaap
author_sort Pronk, Thomas
collection PubMed
description Estimating the reliability of cognitive task datasets is commonly done via split-half methods. We review four methods that differ in how the trials are split into parts: a first-second half split, an odd-even trial split, a permutated split, and a Monte Carlo-based split. Additionally, each splitting method could be combined with stratification by task design. These methods are reviewed in terms of the degree to which they are confounded with four effects that may occur in cognitive tasks: effects of time, task design, trial sampling, and non-linear scoring. Based on the theoretical review, we recommend Monte Carlo splitting (possibly in combination with stratification by task design) as being the most robust method with respect to the four confounds considered. Next, we estimated the reliabilities of the main outcome variables from four cognitive task datasets, each (typically) scored with a different non-linear algorithm, by systematically applying each splitting method. Differences between methods were interpreted in terms of confounding effects inflating or attenuating reliability estimates. For three task datasets, our findings were consistent with our model of confounding effects. Evidence for confounding effects was strong for time and task design and weak for non-linear scoring. When confounding effects occurred, they attenuated reliability estimates. For one task dataset, findings were inconsistent with our model but they may offer indicators for assessing whether a split-half reliability estimate is appropriate. Additionally, we make suggestions on further research of reliability estimation, supported by a compendium R package that implements each of the splitting methods reviewed here.
format Online
Article
Text
id pubmed-8858277
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher Springer US
record_format MEDLINE/PubMed
spelling pubmed-88582772022-02-23 Methods to split cognitive task data for estimating split-half reliability: A comprehensive review and systematic assessment Pronk, Thomas Molenaar, Dylan Wiers, Reinout W. Murre, Jaap Psychon Bull Rev Theoretical/Review Estimating the reliability of cognitive task datasets is commonly done via split-half methods. We review four methods that differ in how the trials are split into parts: a first-second half split, an odd-even trial split, a permutated split, and a Monte Carlo-based split. Additionally, each splitting method could be combined with stratification by task design. These methods are reviewed in terms of the degree to which they are confounded with four effects that may occur in cognitive tasks: effects of time, task design, trial sampling, and non-linear scoring. Based on the theoretical review, we recommend Monte Carlo splitting (possibly in combination with stratification by task design) as being the most robust method with respect to the four confounds considered. Next, we estimated the reliabilities of the main outcome variables from four cognitive task datasets, each (typically) scored with a different non-linear algorithm, by systematically applying each splitting method. Differences between methods were interpreted in terms of confounding effects inflating or attenuating reliability estimates. For three task datasets, our findings were consistent with our model of confounding effects. Evidence for confounding effects was strong for time and task design and weak for non-linear scoring. When confounding effects occurred, they attenuated reliability estimates. For one task dataset, findings were inconsistent with our model but they may offer indicators for assessing whether a split-half reliability estimate is appropriate. Additionally, we make suggestions on further research of reliability estimation, supported by a compendium R package that implements each of the splitting methods reviewed here. Springer US 2021-06-07 2022 /pmc/articles/PMC8858277/ /pubmed/34100223 http://dx.doi.org/10.3758/s13423-021-01948-3 Text en © The Author(s) 2021 https://creativecommons.org/licenses/by/4.0/Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) .
spellingShingle Theoretical/Review
Pronk, Thomas
Molenaar, Dylan
Wiers, Reinout W.
Murre, Jaap
Methods to split cognitive task data for estimating split-half reliability: A comprehensive review and systematic assessment
title Methods to split cognitive task data for estimating split-half reliability: A comprehensive review and systematic assessment
title_full Methods to split cognitive task data for estimating split-half reliability: A comprehensive review and systematic assessment
title_fullStr Methods to split cognitive task data for estimating split-half reliability: A comprehensive review and systematic assessment
title_full_unstemmed Methods to split cognitive task data for estimating split-half reliability: A comprehensive review and systematic assessment
title_short Methods to split cognitive task data for estimating split-half reliability: A comprehensive review and systematic assessment
title_sort methods to split cognitive task data for estimating split-half reliability: a comprehensive review and systematic assessment
topic Theoretical/Review
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8858277/
https://www.ncbi.nlm.nih.gov/pubmed/34100223
http://dx.doi.org/10.3758/s13423-021-01948-3
work_keys_str_mv AT pronkthomas methodstosplitcognitivetaskdataforestimatingsplithalfreliabilityacomprehensivereviewandsystematicassessment
AT molenaardylan methodstosplitcognitivetaskdataforestimatingsplithalfreliabilityacomprehensivereviewandsystematicassessment
AT wiersreinoutw methodstosplitcognitivetaskdataforestimatingsplithalfreliabilityacomprehensivereviewandsystematicassessment
AT murrejaap methodstosplitcognitivetaskdataforestimatingsplithalfreliabilityacomprehensivereviewandsystematicassessment