Cargando…
Estimating meaningful thresholds for multi-item questionnaires using item response theory
PURPOSE: Meaningful thresholds are needed to interpret patient-reported outcome measure (PROM) results. This paper introduces a new method, based on item response theory (IRT), to estimate such thresholds. The performance of the method is examined in simulated datasets and two real datasets, and com...
Autores principales: | , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Springer International Publishing
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10172229/ https://www.ncbi.nlm.nih.gov/pubmed/36780033 http://dx.doi.org/10.1007/s11136-023-03355-8 |
_version_ | 1785039577970376704 |
---|---|
author | Terluin, Berend Koopman, Jaimy E. Hoogendam, Lisa Griffiths, Pip Terwee, Caroline B. Bjorner, Jakob B. |
author_facet | Terluin, Berend Koopman, Jaimy E. Hoogendam, Lisa Griffiths, Pip Terwee, Caroline B. Bjorner, Jakob B. |
author_sort | Terluin, Berend |
collection | PubMed |
description | PURPOSE: Meaningful thresholds are needed to interpret patient-reported outcome measure (PROM) results. This paper introduces a new method, based on item response theory (IRT), to estimate such thresholds. The performance of the method is examined in simulated datasets and two real datasets, and compared with other methods. METHODS: The IRT method involves fitting an IRT model to the PROM items and an anchor item indicating the criterion state of interest. The difficulty parameter of the anchor item represents the meaningful threshold on the latent trait. The latent threshold is then linked to the corresponding expected PROM score. We simulated 4500 item response datasets to a 10-item PROM, and an anchor item. The datasets varied with respect to the mean and standard deviation of the latent trait, and the reliability of the anchor item. The real datasets consisted of a depression scale with a clinical depression diagnosis as anchor variable and a pain scale with a patient acceptable symptom state (PASS) question as anchor variable. RESULTS: The new IRT method recovered the true thresholds accurately across the simulated datasets. The other methods, except one, produced biased threshold estimates if the state prevalence was smaller or greater than 0.5. The adjusted predictive modeling method matched the new IRT method (also in the real datasets) but showed some residual bias if the prevalence was smaller than 0.3 or greater than 0.7. CONCLUSIONS: The new IRT method perfectly recovers meaningful (interpretational) thresholds for multi-item questionnaires, provided that the data satisfy the assumptions for IRT analysis. SUPPLEMENTARY INFORMATION: The online version contains supplementary material available at 10.1007/s11136-023-03355-8. |
format | Online Article Text |
id | pubmed-10172229 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Springer International Publishing |
record_format | MEDLINE/PubMed |
spelling | pubmed-101722292023-05-12 Estimating meaningful thresholds for multi-item questionnaires using item response theory Terluin, Berend Koopman, Jaimy E. Hoogendam, Lisa Griffiths, Pip Terwee, Caroline B. Bjorner, Jakob B. Qual Life Res Article PURPOSE: Meaningful thresholds are needed to interpret patient-reported outcome measure (PROM) results. This paper introduces a new method, based on item response theory (IRT), to estimate such thresholds. The performance of the method is examined in simulated datasets and two real datasets, and compared with other methods. METHODS: The IRT method involves fitting an IRT model to the PROM items and an anchor item indicating the criterion state of interest. The difficulty parameter of the anchor item represents the meaningful threshold on the latent trait. The latent threshold is then linked to the corresponding expected PROM score. We simulated 4500 item response datasets to a 10-item PROM, and an anchor item. The datasets varied with respect to the mean and standard deviation of the latent trait, and the reliability of the anchor item. The real datasets consisted of a depression scale with a clinical depression diagnosis as anchor variable and a pain scale with a patient acceptable symptom state (PASS) question as anchor variable. RESULTS: The new IRT method recovered the true thresholds accurately across the simulated datasets. The other methods, except one, produced biased threshold estimates if the state prevalence was smaller or greater than 0.5. The adjusted predictive modeling method matched the new IRT method (also in the real datasets) but showed some residual bias if the prevalence was smaller than 0.3 or greater than 0.7. CONCLUSIONS: The new IRT method perfectly recovers meaningful (interpretational) thresholds for multi-item questionnaires, provided that the data satisfy the assumptions for IRT analysis. SUPPLEMENTARY INFORMATION: The online version contains supplementary material available at 10.1007/s11136-023-03355-8. Springer International Publishing 2023-02-13 2023 /pmc/articles/PMC10172229/ /pubmed/36780033 http://dx.doi.org/10.1007/s11136-023-03355-8 Text en © The Author(s) 2023 https://creativecommons.org/licenses/by/4.0/Open AccessThis article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) . |
spellingShingle | Article Terluin, Berend Koopman, Jaimy E. Hoogendam, Lisa Griffiths, Pip Terwee, Caroline B. Bjorner, Jakob B. Estimating meaningful thresholds for multi-item questionnaires using item response theory |
title | Estimating meaningful thresholds for multi-item questionnaires using item response theory |
title_full | Estimating meaningful thresholds for multi-item questionnaires using item response theory |
title_fullStr | Estimating meaningful thresholds for multi-item questionnaires using item response theory |
title_full_unstemmed | Estimating meaningful thresholds for multi-item questionnaires using item response theory |
title_short | Estimating meaningful thresholds for multi-item questionnaires using item response theory |
title_sort | estimating meaningful thresholds for multi-item questionnaires using item response theory |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10172229/ https://www.ncbi.nlm.nih.gov/pubmed/36780033 http://dx.doi.org/10.1007/s11136-023-03355-8 |
work_keys_str_mv | AT terluinberend estimatingmeaningfulthresholdsformultiitemquestionnairesusingitemresponsetheory AT koopmanjaimye estimatingmeaningfulthresholdsformultiitemquestionnairesusingitemresponsetheory AT hoogendamlisa estimatingmeaningfulthresholdsformultiitemquestionnairesusingitemresponsetheory AT griffithspip estimatingmeaningfulthresholdsformultiitemquestionnairesusingitemresponsetheory AT terweecarolineb estimatingmeaningfulthresholdsformultiitemquestionnairesusingitemresponsetheory AT bjornerjakobb estimatingmeaningfulthresholdsformultiitemquestionnairesusingitemresponsetheory |