Cargando…
Methods to split cognitive task data for estimating split-half reliability: A comprehensive review and systematic assessment
Estimating the reliability of cognitive task datasets is commonly done via split-half methods. We review four methods that differ in how the trials are split into parts: a first-second half split, an odd-even trial split, a permutated split, and a Monte Carlo-based split. Additionally, each splittin...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Springer US
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8858277/ https://www.ncbi.nlm.nih.gov/pubmed/34100223 http://dx.doi.org/10.3758/s13423-021-01948-3 |
_version_ | 1784654216148549632 |
---|---|
author | Pronk, Thomas Molenaar, Dylan Wiers, Reinout W. Murre, Jaap |
author_facet | Pronk, Thomas Molenaar, Dylan Wiers, Reinout W. Murre, Jaap |
author_sort | Pronk, Thomas |
collection | PubMed |
description | Estimating the reliability of cognitive task datasets is commonly done via split-half methods. We review four methods that differ in how the trials are split into parts: a first-second half split, an odd-even trial split, a permutated split, and a Monte Carlo-based split. Additionally, each splitting method could be combined with stratification by task design. These methods are reviewed in terms of the degree to which they are confounded with four effects that may occur in cognitive tasks: effects of time, task design, trial sampling, and non-linear scoring. Based on the theoretical review, we recommend Monte Carlo splitting (possibly in combination with stratification by task design) as being the most robust method with respect to the four confounds considered. Next, we estimated the reliabilities of the main outcome variables from four cognitive task datasets, each (typically) scored with a different non-linear algorithm, by systematically applying each splitting method. Differences between methods were interpreted in terms of confounding effects inflating or attenuating reliability estimates. For three task datasets, our findings were consistent with our model of confounding effects. Evidence for confounding effects was strong for time and task design and weak for non-linear scoring. When confounding effects occurred, they attenuated reliability estimates. For one task dataset, findings were inconsistent with our model but they may offer indicators for assessing whether a split-half reliability estimate is appropriate. Additionally, we make suggestions on further research of reliability estimation, supported by a compendium R package that implements each of the splitting methods reviewed here. |
format | Online Article Text |
id | pubmed-8858277 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | Springer US |
record_format | MEDLINE/PubMed |
spelling | pubmed-88582772022-02-23 Methods to split cognitive task data for estimating split-half reliability: A comprehensive review and systematic assessment Pronk, Thomas Molenaar, Dylan Wiers, Reinout W. Murre, Jaap Psychon Bull Rev Theoretical/Review Estimating the reliability of cognitive task datasets is commonly done via split-half methods. We review four methods that differ in how the trials are split into parts: a first-second half split, an odd-even trial split, a permutated split, and a Monte Carlo-based split. Additionally, each splitting method could be combined with stratification by task design. These methods are reviewed in terms of the degree to which they are confounded with four effects that may occur in cognitive tasks: effects of time, task design, trial sampling, and non-linear scoring. Based on the theoretical review, we recommend Monte Carlo splitting (possibly in combination with stratification by task design) as being the most robust method with respect to the four confounds considered. Next, we estimated the reliabilities of the main outcome variables from four cognitive task datasets, each (typically) scored with a different non-linear algorithm, by systematically applying each splitting method. Differences between methods were interpreted in terms of confounding effects inflating or attenuating reliability estimates. For three task datasets, our findings were consistent with our model of confounding effects. Evidence for confounding effects was strong for time and task design and weak for non-linear scoring. When confounding effects occurred, they attenuated reliability estimates. For one task dataset, findings were inconsistent with our model but they may offer indicators for assessing whether a split-half reliability estimate is appropriate. Additionally, we make suggestions on further research of reliability estimation, supported by a compendium R package that implements each of the splitting methods reviewed here. Springer US 2021-06-07 2022 /pmc/articles/PMC8858277/ /pubmed/34100223 http://dx.doi.org/10.3758/s13423-021-01948-3 Text en © The Author(s) 2021 https://creativecommons.org/licenses/by/4.0/Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) . |
spellingShingle | Theoretical/Review Pronk, Thomas Molenaar, Dylan Wiers, Reinout W. Murre, Jaap Methods to split cognitive task data for estimating split-half reliability: A comprehensive review and systematic assessment |
title | Methods to split cognitive task data for estimating split-half reliability: A comprehensive review and systematic assessment |
title_full | Methods to split cognitive task data for estimating split-half reliability: A comprehensive review and systematic assessment |
title_fullStr | Methods to split cognitive task data for estimating split-half reliability: A comprehensive review and systematic assessment |
title_full_unstemmed | Methods to split cognitive task data for estimating split-half reliability: A comprehensive review and systematic assessment |
title_short | Methods to split cognitive task data for estimating split-half reliability: A comprehensive review and systematic assessment |
title_sort | methods to split cognitive task data for estimating split-half reliability: a comprehensive review and systematic assessment |
topic | Theoretical/Review |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8858277/ https://www.ncbi.nlm.nih.gov/pubmed/34100223 http://dx.doi.org/10.3758/s13423-021-01948-3 |
work_keys_str_mv | AT pronkthomas methodstosplitcognitivetaskdataforestimatingsplithalfreliabilityacomprehensivereviewandsystematicassessment AT molenaardylan methodstosplitcognitivetaskdataforestimatingsplithalfreliabilityacomprehensivereviewandsystematicassessment AT wiersreinoutw methodstosplitcognitivetaskdataforestimatingsplithalfreliabilityacomprehensivereviewandsystematicassessment AT murrejaap methodstosplitcognitivetaskdataforestimatingsplithalfreliabilityacomprehensivereviewandsystematicassessment |