Cargando…
Item-Score Reliability as a Selection Tool in Test Construction
This study investigates the usefulness of item-score reliability as a criterion for item selection in test construction. Methods MS, λ(6), and CA were investigated as item-assessment methods in item selection and compared to the corrected item-total correlation, which was used as a benchmark. An ide...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2019
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6336834/ https://www.ncbi.nlm.nih.gov/pubmed/30687144 http://dx.doi.org/10.3389/fpsyg.2018.02298 |
_version_ | 1783388127292817408 |
---|---|
author | Zijlmans, Eva A. O. Tijmstra, Jesper van der Ark, L. Andries Sijtsma, Klaas |
author_facet | Zijlmans, Eva A. O. Tijmstra, Jesper van der Ark, L. Andries Sijtsma, Klaas |
author_sort | Zijlmans, Eva A. O. |
collection | PubMed |
description | This study investigates the usefulness of item-score reliability as a criterion for item selection in test construction. Methods MS, λ(6), and CA were investigated as item-assessment methods in item selection and compared to the corrected item-total correlation, which was used as a benchmark. An ideal ordering to add items to the test (bottom-up procedure) or omit items from the test (top-down procedure) was defined based on the population test-score reliability. The orderings the four item-assessment methods produced in samples were compared to the ideal ordering, and the degree of resemblance was expressed by means of Kendall's τ. To investigate the concordance of the orderings across 1,000 replicated samples, Kendall's W was computed for each item-assessment method. The results showed that for both the bottom-up and the top-down procedures, item-assessment method CA and the corrected item-total correlation most closely resembled the ideal ordering. Generally, all item assessment methods resembled the ideal ordering better, and concordance of the orderings was greater, for larger sample sizes, and greater variance of the item discrimination parameters. |
format | Online Article Text |
id | pubmed-6336834 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2019 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-63368342019-01-25 Item-Score Reliability as a Selection Tool in Test Construction Zijlmans, Eva A. O. Tijmstra, Jesper van der Ark, L. Andries Sijtsma, Klaas Front Psychol Psychology This study investigates the usefulness of item-score reliability as a criterion for item selection in test construction. Methods MS, λ(6), and CA were investigated as item-assessment methods in item selection and compared to the corrected item-total correlation, which was used as a benchmark. An ideal ordering to add items to the test (bottom-up procedure) or omit items from the test (top-down procedure) was defined based on the population test-score reliability. The orderings the four item-assessment methods produced in samples were compared to the ideal ordering, and the degree of resemblance was expressed by means of Kendall's τ. To investigate the concordance of the orderings across 1,000 replicated samples, Kendall's W was computed for each item-assessment method. The results showed that for both the bottom-up and the top-down procedures, item-assessment method CA and the corrected item-total correlation most closely resembled the ideal ordering. Generally, all item assessment methods resembled the ideal ordering better, and concordance of the orderings was greater, for larger sample sizes, and greater variance of the item discrimination parameters. Frontiers Media S.A. 2019-01-11 /pmc/articles/PMC6336834/ /pubmed/30687144 http://dx.doi.org/10.3389/fpsyg.2018.02298 Text en Copyright © 2019 Zijlmans, Tijmstra, van der Ark and Sijtsma. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Psychology Zijlmans, Eva A. O. Tijmstra, Jesper van der Ark, L. Andries Sijtsma, Klaas Item-Score Reliability as a Selection Tool in Test Construction |
title | Item-Score Reliability as a Selection Tool in Test Construction |
title_full | Item-Score Reliability as a Selection Tool in Test Construction |
title_fullStr | Item-Score Reliability as a Selection Tool in Test Construction |
title_full_unstemmed | Item-Score Reliability as a Selection Tool in Test Construction |
title_short | Item-Score Reliability as a Selection Tool in Test Construction |
title_sort | item-score reliability as a selection tool in test construction |
topic | Psychology |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6336834/ https://www.ncbi.nlm.nih.gov/pubmed/30687144 http://dx.doi.org/10.3389/fpsyg.2018.02298 |
work_keys_str_mv | AT zijlmansevaao itemscorereliabilityasaselectiontoolintestconstruction AT tijmstrajesper itemscorereliabilityasaselectiontoolintestconstruction AT vanderarklandries itemscorereliabilityasaselectiontoolintestconstruction AT sijtsmaklaas itemscorereliabilityasaselectiontoolintestconstruction |