Cargando…

Conducting Online Expert panels: a feasibility and experimental replicability study

BACKGROUND: This paper has two goals. First, we explore the feasibility of conducting online expert panels to facilitate consensus finding among a large number of geographically distributed stakeholders. Second, we test the replicability of panel findings across four panels of different size. METHOD...

Descripción completa

Detalles Bibliográficos
Autores principales: Khodyakov, Dmitry, Hempel, Susanne, Rubenstein, Lisa, Shekelle, Paul, Foy, Robbie, Salem-Schatz, Susanne, O'Neill, Sean, Danz, Margie, Dalal, Siddhartha
Formato: Online Artículo Texto
Lenguaje:English
Publicado: BioMed Central 2011
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3313865/
https://www.ncbi.nlm.nih.gov/pubmed/22196011
http://dx.doi.org/10.1186/1471-2288-11-174
_version_ 1782228039683276800
author Khodyakov, Dmitry
Hempel, Susanne
Rubenstein, Lisa
Shekelle, Paul
Foy, Robbie
Salem-Schatz, Susanne
O'Neill, Sean
Danz, Margie
Dalal, Siddhartha
author_facet Khodyakov, Dmitry
Hempel, Susanne
Rubenstein, Lisa
Shekelle, Paul
Foy, Robbie
Salem-Schatz, Susanne
O'Neill, Sean
Danz, Margie
Dalal, Siddhartha
author_sort Khodyakov, Dmitry
collection PubMed
description BACKGROUND: This paper has two goals. First, we explore the feasibility of conducting online expert panels to facilitate consensus finding among a large number of geographically distributed stakeholders. Second, we test the replicability of panel findings across four panels of different size. METHOD: We engaged 119 panelists in an iterative process to identify definitional features of Continuous Quality Improvement (CQI). We conducted four parallel online panels of different size through three one-week phases by using the RAND's ExpertLens process. In Phase I, participants rated potentially definitional CQI features. In Phase II, they discussed rating results online, using asynchronous, anonymous discussion boards. In Phase III, panelists re-rated Phase I features and reported on their experiences as participants. RESULTS: 66% of invited experts participated in all three phases. 62% of Phase I participants contributed to Phase II discussions and 87% of them completed Phase III. Panel disagreement, measured by the mean absolute deviation from the median (MAD-M), decreased after group feedback and discussion in 36 out of 43 judgments about CQI features. Agreement between the four panels after Phase III was fair (four-way kappa = 0.36); they agreed on the status of five out of eleven CQI features. Results of the post-completion survey suggest that participants were generally satisfied with the online process. Compared to participants in smaller panels, those in larger panels were more likely to agree that they had debated each others' view points. CONCLUSION: It is feasible to conduct online expert panels intended to facilitate consensus finding among geographically distributed participants. The online approach may be practical for engaging large and diverse groups of stakeholders around a range of health services research topics and can help conduct multiple parallel panels to test for the reproducibility of panel conclusions.
format Online
Article
Text
id pubmed-3313865
institution National Center for Biotechnology Information
language English
publishDate 2011
publisher BioMed Central
record_format MEDLINE/PubMed
spelling pubmed-33138652012-03-28 Conducting Online Expert panels: a feasibility and experimental replicability study Khodyakov, Dmitry Hempel, Susanne Rubenstein, Lisa Shekelle, Paul Foy, Robbie Salem-Schatz, Susanne O'Neill, Sean Danz, Margie Dalal, Siddhartha BMC Med Res Methodol Research Article BACKGROUND: This paper has two goals. First, we explore the feasibility of conducting online expert panels to facilitate consensus finding among a large number of geographically distributed stakeholders. Second, we test the replicability of panel findings across four panels of different size. METHOD: We engaged 119 panelists in an iterative process to identify definitional features of Continuous Quality Improvement (CQI). We conducted four parallel online panels of different size through three one-week phases by using the RAND's ExpertLens process. In Phase I, participants rated potentially definitional CQI features. In Phase II, they discussed rating results online, using asynchronous, anonymous discussion boards. In Phase III, panelists re-rated Phase I features and reported on their experiences as participants. RESULTS: 66% of invited experts participated in all three phases. 62% of Phase I participants contributed to Phase II discussions and 87% of them completed Phase III. Panel disagreement, measured by the mean absolute deviation from the median (MAD-M), decreased after group feedback and discussion in 36 out of 43 judgments about CQI features. Agreement between the four panels after Phase III was fair (four-way kappa = 0.36); they agreed on the status of five out of eleven CQI features. Results of the post-completion survey suggest that participants were generally satisfied with the online process. Compared to participants in smaller panels, those in larger panels were more likely to agree that they had debated each others' view points. CONCLUSION: It is feasible to conduct online expert panels intended to facilitate consensus finding among geographically distributed participants. The online approach may be practical for engaging large and diverse groups of stakeholders around a range of health services research topics and can help conduct multiple parallel panels to test for the reproducibility of panel conclusions. BioMed Central 2011-12-23 /pmc/articles/PMC3313865/ /pubmed/22196011 http://dx.doi.org/10.1186/1471-2288-11-174 Text en Copyright ©2011 Khodyakov et al; licensee BioMed Central Ltd. http://creativecommons.org/licenses/by/2.0 This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
spellingShingle Research Article
Khodyakov, Dmitry
Hempel, Susanne
Rubenstein, Lisa
Shekelle, Paul
Foy, Robbie
Salem-Schatz, Susanne
O'Neill, Sean
Danz, Margie
Dalal, Siddhartha
Conducting Online Expert panels: a feasibility and experimental replicability study
title Conducting Online Expert panels: a feasibility and experimental replicability study
title_full Conducting Online Expert panels: a feasibility and experimental replicability study
title_fullStr Conducting Online Expert panels: a feasibility and experimental replicability study
title_full_unstemmed Conducting Online Expert panels: a feasibility and experimental replicability study
title_short Conducting Online Expert panels: a feasibility and experimental replicability study
title_sort conducting online expert panels: a feasibility and experimental replicability study
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3313865/
https://www.ncbi.nlm.nih.gov/pubmed/22196011
http://dx.doi.org/10.1186/1471-2288-11-174
work_keys_str_mv AT khodyakovdmitry conductingonlineexpertpanelsafeasibilityandexperimentalreplicabilitystudy
AT hempelsusanne conductingonlineexpertpanelsafeasibilityandexperimentalreplicabilitystudy
AT rubensteinlisa conductingonlineexpertpanelsafeasibilityandexperimentalreplicabilitystudy
AT shekellepaul conductingonlineexpertpanelsafeasibilityandexperimentalreplicabilitystudy
AT foyrobbie conductingonlineexpertpanelsafeasibilityandexperimentalreplicabilitystudy
AT salemschatzsusanne conductingonlineexpertpanelsafeasibilityandexperimentalreplicabilitystudy
AT oneillsean conductingonlineexpertpanelsafeasibilityandexperimentalreplicabilitystudy
AT danzmargie conductingonlineexpertpanelsafeasibilityandexperimentalreplicabilitystudy
AT dalalsiddhartha conductingonlineexpertpanelsafeasibilityandexperimentalreplicabilitystudy