Cargando…
Direct Observation Assessment of Milestones: Problems with Reliability
INTRODUCTION: Emergency medicine (EM) milestones are used to assess residents’ progress. While some milestone validity evidence exists, there is a lack of standardized tools available to reliably assess residents. Inherent to this is a concern that we may not be truly measuring what we intend to ass...
Autores principales: | , , , , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Department of Emergency Medicine, University of California, Irvine School of Medicine
2015
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4651585/ https://www.ncbi.nlm.nih.gov/pubmed/26594281 http://dx.doi.org/10.5811/westjem.2015.9.27270 |
_version_ | 1782401656870141952 |
---|---|
author | Schott, Meghan Kedia, Raashee Promes, Susan B. Swoboda, Thomas O’Rourke, Kevin Green, Walter Liu, Rachel Stansfield, Brent Santen, Sally A. |
author_facet | Schott, Meghan Kedia, Raashee Promes, Susan B. Swoboda, Thomas O’Rourke, Kevin Green, Walter Liu, Rachel Stansfield, Brent Santen, Sally A. |
author_sort | Schott, Meghan |
collection | PubMed |
description | INTRODUCTION: Emergency medicine (EM) milestones are used to assess residents’ progress. While some milestone validity evidence exists, there is a lack of standardized tools available to reliably assess residents. Inherent to this is a concern that we may not be truly measuring what we intend to assess. The purpose of this study was to design a direct observation milestone assessment instrument supported by validity and reliability evidence. In addition, such a tool would further lend validity evidence to the EM milestones by demonstrating their accurate measurement. METHODS: This was a multi-center, prospective, observational validity study conducted at eight institutions. The Critical Care Direct Observation Tool (CDOT) was created to assess EM residents during resuscitations. This tool was designed using a modified Delphi method focused on content, response process, and internal structure validity. Paying special attention to content validity, the CDOT was developed by an expert panel, maintaining the use of the EM milestone wording. We built response process and internal consistency by piloting and revising the instrument. Raters were faculty who routinely assess residents on the milestones. A brief training video on utilization of the instrument was completed by all. Raters used the CDOT to assess simulated videos of three residents at different stages of training in a critical care scenario. We measured reliability using Fleiss’ kappa and interclass correlations. RESULTS: Two versions of the CDOT were used: one used the milestone levels as global rating scales with anchors, and the second reflected a current trend of a checklist response system. Although the raters who used the CDOT routinely rate residents in their practice, they did not score the residents’ performances in the videos comparably, which led to poor reliability. The Fleiss’ kappa of each of the items measured on both versions of the CDOT was near zero. CONCLUSION: The validity and reliability of the current EM milestone assessment tools have yet to be determined. This study is a rigorous attempt to collect validity evidence in the development of a direct observation assessment instrument. However, despite strict attention to validity evidence, inter-rater reliability was low. The potential sources of reducible variance include rater- and instrument-based error. Based on this study, there may be concerns for the reliability of other EM milestone assessment tools that are currently in use. |
format | Online Article Text |
id | pubmed-4651585 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2015 |
publisher | Department of Emergency Medicine, University of California, Irvine School of Medicine |
record_format | MEDLINE/PubMed |
spelling | pubmed-46515852015-11-20 Direct Observation Assessment of Milestones: Problems with Reliability Schott, Meghan Kedia, Raashee Promes, Susan B. Swoboda, Thomas O’Rourke, Kevin Green, Walter Liu, Rachel Stansfield, Brent Santen, Sally A. West J Emerg Med Educational Research and Practice INTRODUCTION: Emergency medicine (EM) milestones are used to assess residents’ progress. While some milestone validity evidence exists, there is a lack of standardized tools available to reliably assess residents. Inherent to this is a concern that we may not be truly measuring what we intend to assess. The purpose of this study was to design a direct observation milestone assessment instrument supported by validity and reliability evidence. In addition, such a tool would further lend validity evidence to the EM milestones by demonstrating their accurate measurement. METHODS: This was a multi-center, prospective, observational validity study conducted at eight institutions. The Critical Care Direct Observation Tool (CDOT) was created to assess EM residents during resuscitations. This tool was designed using a modified Delphi method focused on content, response process, and internal structure validity. Paying special attention to content validity, the CDOT was developed by an expert panel, maintaining the use of the EM milestone wording. We built response process and internal consistency by piloting and revising the instrument. Raters were faculty who routinely assess residents on the milestones. A brief training video on utilization of the instrument was completed by all. Raters used the CDOT to assess simulated videos of three residents at different stages of training in a critical care scenario. We measured reliability using Fleiss’ kappa and interclass correlations. RESULTS: Two versions of the CDOT were used: one used the milestone levels as global rating scales with anchors, and the second reflected a current trend of a checklist response system. Although the raters who used the CDOT routinely rate residents in their practice, they did not score the residents’ performances in the videos comparably, which led to poor reliability. The Fleiss’ kappa of each of the items measured on both versions of the CDOT was near zero. CONCLUSION: The validity and reliability of the current EM milestone assessment tools have yet to be determined. This study is a rigorous attempt to collect validity evidence in the development of a direct observation assessment instrument. However, despite strict attention to validity evidence, inter-rater reliability was low. The potential sources of reducible variance include rater- and instrument-based error. Based on this study, there may be concerns for the reliability of other EM milestone assessment tools that are currently in use. Department of Emergency Medicine, University of California, Irvine School of Medicine 2015-11 2015-10-22 /pmc/articles/PMC4651585/ /pubmed/26594281 http://dx.doi.org/10.5811/westjem.2015.9.27270 Text en Copyright © 2015 Schott et al. http://creativecommons.org/licenses/by/4.0/ This is an open access article distributed in accordance with the terms of the Creative Commons Attribution (CC BY 4.0) License. See: http://creativecommons.org/licenses/by/4.0/ |
spellingShingle | Educational Research and Practice Schott, Meghan Kedia, Raashee Promes, Susan B. Swoboda, Thomas O’Rourke, Kevin Green, Walter Liu, Rachel Stansfield, Brent Santen, Sally A. Direct Observation Assessment of Milestones: Problems with Reliability |
title | Direct Observation Assessment of Milestones: Problems with Reliability |
title_full | Direct Observation Assessment of Milestones: Problems with Reliability |
title_fullStr | Direct Observation Assessment of Milestones: Problems with Reliability |
title_full_unstemmed | Direct Observation Assessment of Milestones: Problems with Reliability |
title_short | Direct Observation Assessment of Milestones: Problems with Reliability |
title_sort | direct observation assessment of milestones: problems with reliability |
topic | Educational Research and Practice |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4651585/ https://www.ncbi.nlm.nih.gov/pubmed/26594281 http://dx.doi.org/10.5811/westjem.2015.9.27270 |
work_keys_str_mv | AT schottmeghan directobservationassessmentofmilestonesproblemswithreliability AT kediaraashee directobservationassessmentofmilestonesproblemswithreliability AT promessusanb directobservationassessmentofmilestonesproblemswithreliability AT swobodathomas directobservationassessmentofmilestonesproblemswithreliability AT orourkekevin directobservationassessmentofmilestonesproblemswithreliability AT greenwalter directobservationassessmentofmilestonesproblemswithreliability AT liurachel directobservationassessmentofmilestonesproblemswithreliability AT stansfieldbrent directobservationassessmentofmilestonesproblemswithreliability AT santensallya directobservationassessmentofmilestonesproblemswithreliability |