Cargando…

Item-Score Reliability as a Selection Tool in Test Construction

This study investigates the usefulness of item-score reliability as a criterion for item selection in test construction. Methods MS, λ(6), and CA were investigated as item-assessment methods in item selection and compared to the corrected item-total correlation, which was used as a benchmark. An ide...

Descripción completa

Detalles Bibliográficos
Autores principales: Zijlmans, Eva A. O., Tijmstra, Jesper, van der Ark, L. Andries, Sijtsma, Klaas
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2019
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6336834/
https://www.ncbi.nlm.nih.gov/pubmed/30687144
http://dx.doi.org/10.3389/fpsyg.2018.02298
_version_ 1783388127292817408
author Zijlmans, Eva A. O.
Tijmstra, Jesper
van der Ark, L. Andries
Sijtsma, Klaas
author_facet Zijlmans, Eva A. O.
Tijmstra, Jesper
van der Ark, L. Andries
Sijtsma, Klaas
author_sort Zijlmans, Eva A. O.
collection PubMed
description This study investigates the usefulness of item-score reliability as a criterion for item selection in test construction. Methods MS, λ(6), and CA were investigated as item-assessment methods in item selection and compared to the corrected item-total correlation, which was used as a benchmark. An ideal ordering to add items to the test (bottom-up procedure) or omit items from the test (top-down procedure) was defined based on the population test-score reliability. The orderings the four item-assessment methods produced in samples were compared to the ideal ordering, and the degree of resemblance was expressed by means of Kendall's τ. To investigate the concordance of the orderings across 1,000 replicated samples, Kendall's W was computed for each item-assessment method. The results showed that for both the bottom-up and the top-down procedures, item-assessment method CA and the corrected item-total correlation most closely resembled the ideal ordering. Generally, all item assessment methods resembled the ideal ordering better, and concordance of the orderings was greater, for larger sample sizes, and greater variance of the item discrimination parameters.
format Online
Article
Text
id pubmed-6336834
institution National Center for Biotechnology Information
language English
publishDate 2019
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-63368342019-01-25 Item-Score Reliability as a Selection Tool in Test Construction Zijlmans, Eva A. O. Tijmstra, Jesper van der Ark, L. Andries Sijtsma, Klaas Front Psychol Psychology This study investigates the usefulness of item-score reliability as a criterion for item selection in test construction. Methods MS, λ(6), and CA were investigated as item-assessment methods in item selection and compared to the corrected item-total correlation, which was used as a benchmark. An ideal ordering to add items to the test (bottom-up procedure) or omit items from the test (top-down procedure) was defined based on the population test-score reliability. The orderings the four item-assessment methods produced in samples were compared to the ideal ordering, and the degree of resemblance was expressed by means of Kendall's τ. To investigate the concordance of the orderings across 1,000 replicated samples, Kendall's W was computed for each item-assessment method. The results showed that for both the bottom-up and the top-down procedures, item-assessment method CA and the corrected item-total correlation most closely resembled the ideal ordering. Generally, all item assessment methods resembled the ideal ordering better, and concordance of the orderings was greater, for larger sample sizes, and greater variance of the item discrimination parameters. Frontiers Media S.A. 2019-01-11 /pmc/articles/PMC6336834/ /pubmed/30687144 http://dx.doi.org/10.3389/fpsyg.2018.02298 Text en Copyright © 2019 Zijlmans, Tijmstra, van der Ark and Sijtsma. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Psychology
Zijlmans, Eva A. O.
Tijmstra, Jesper
van der Ark, L. Andries
Sijtsma, Klaas
Item-Score Reliability as a Selection Tool in Test Construction
title Item-Score Reliability as a Selection Tool in Test Construction
title_full Item-Score Reliability as a Selection Tool in Test Construction
title_fullStr Item-Score Reliability as a Selection Tool in Test Construction
title_full_unstemmed Item-Score Reliability as a Selection Tool in Test Construction
title_short Item-Score Reliability as a Selection Tool in Test Construction
title_sort item-score reliability as a selection tool in test construction
topic Psychology
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6336834/
https://www.ncbi.nlm.nih.gov/pubmed/30687144
http://dx.doi.org/10.3389/fpsyg.2018.02298
work_keys_str_mv AT zijlmansevaao itemscorereliabilityasaselectiontoolintestconstruction
AT tijmstrajesper itemscorereliabilityasaselectiontoolintestconstruction
AT vanderarklandries itemscorereliabilityasaselectiontoolintestconstruction
AT sijtsmaklaas itemscorereliabilityasaselectiontoolintestconstruction