Cargando…
Is It All About the Form? Norm- vs Criterion-Referenced Ratings and Faculty Inter-Rater Reliability
Background: Little research to date has examined the quality of data obtained from resident performance evaluations. This study sought to address this need and compared inter-rater reliability obtained from norm-referenced and criterion-referenced evaluation scaling approaches for faculty completing...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Academic Division of Ochsner Clinic Foundation
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10498947/ https://www.ncbi.nlm.nih.gov/pubmed/37711480 http://dx.doi.org/10.31486/toj.23.0014 |
_version_ | 1785105626992476160 |
---|---|
author | Scielzo, Shannon A. Abdelfattah, Kareem Ryder, Hilary F. |
author_facet | Scielzo, Shannon A. Abdelfattah, Kareem Ryder, Hilary F. |
author_sort | Scielzo, Shannon A. |
collection | PubMed |
description | Background: Little research to date has examined the quality of data obtained from resident performance evaluations. This study sought to address this need and compared inter-rater reliability obtained from norm-referenced and criterion-referenced evaluation scaling approaches for faculty completing resident performance evaluations. Methods: Resident performance evaluation data were examined from 2 institutions (3 programs, 2 internal medicine and 1 surgery; 426 residents in total), with 4 evaluation forms: 2 criterion-referenced (1 with an additional norm-referenced item) and 2 norm-referenced. Faculty inter-rater reliability was calculated with intraclass correlation coefficients (ICCs) (1,10) for each competency area within the form. ICCs were transformed to z-scores, and 95% CIs were computed. Reliabilities for each evaluation form and competency, averages within competency, and averages within scaling type were examined. Results: Inter-rater reliability averages were higher for all competencies that used criterion-referenced scaling relative to those that used norm-referenced scaling. Aggregate scores of all independent categories (competencies and the items assessing overall competence) for criterion-referenced scaling demonstrated higher reliability (z=1.37, CI 1.26-1.48) than norm-referenced scaling (z=0.88, CI 0.77-0.99). Moreover, examination of the distributions of composite scores (average of all competencies and raters for each individual being rated) suggested that the criterion-referenced evaluations better represented the performance continuum. Conclusion: Criterion-referenced evaluation approaches appear to provide superior inter-rater reliability relative to norm-referenced evaluation scaling approaches. Although more research is needed to identify resident evaluation best practices, using criterion-referenced scaling may provide more valid data than norm-referenced scaling. |
format | Online Article Text |
id | pubmed-10498947 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Academic Division of Ochsner Clinic Foundation |
record_format | MEDLINE/PubMed |
spelling | pubmed-104989472023-09-14 Is It All About the Form? Norm- vs Criterion-Referenced Ratings and Faculty Inter-Rater Reliability Scielzo, Shannon A. Abdelfattah, Kareem Ryder, Hilary F. Ochsner J Original Research Background: Little research to date has examined the quality of data obtained from resident performance evaluations. This study sought to address this need and compared inter-rater reliability obtained from norm-referenced and criterion-referenced evaluation scaling approaches for faculty completing resident performance evaluations. Methods: Resident performance evaluation data were examined from 2 institutions (3 programs, 2 internal medicine and 1 surgery; 426 residents in total), with 4 evaluation forms: 2 criterion-referenced (1 with an additional norm-referenced item) and 2 norm-referenced. Faculty inter-rater reliability was calculated with intraclass correlation coefficients (ICCs) (1,10) for each competency area within the form. ICCs were transformed to z-scores, and 95% CIs were computed. Reliabilities for each evaluation form and competency, averages within competency, and averages within scaling type were examined. Results: Inter-rater reliability averages were higher for all competencies that used criterion-referenced scaling relative to those that used norm-referenced scaling. Aggregate scores of all independent categories (competencies and the items assessing overall competence) for criterion-referenced scaling demonstrated higher reliability (z=1.37, CI 1.26-1.48) than norm-referenced scaling (z=0.88, CI 0.77-0.99). Moreover, examination of the distributions of composite scores (average of all competencies and raters for each individual being rated) suggested that the criterion-referenced evaluations better represented the performance continuum. Conclusion: Criterion-referenced evaluation approaches appear to provide superior inter-rater reliability relative to norm-referenced evaluation scaling approaches. Although more research is needed to identify resident evaluation best practices, using criterion-referenced scaling may provide more valid data than norm-referenced scaling. Academic Division of Ochsner Clinic Foundation 2023 2023 /pmc/articles/PMC10498947/ /pubmed/37711480 http://dx.doi.org/10.31486/toj.23.0014 Text en ©2023 by the author(s); Creative Commons Attribution License (CC BY) https://creativecommons.org/licenses/by/4.0/©2023 by the author(s); licensee Ochsner Journal, Ochsner Clinic Foundation, New Orleans, LA. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (creativecommons.org/licenses/by/4.0/legalcode) that permits unrestricted use, distribution, and reproduction in any medium, provided the original author(s) and source are credited. |
spellingShingle | Original Research Scielzo, Shannon A. Abdelfattah, Kareem Ryder, Hilary F. Is It All About the Form? Norm- vs Criterion-Referenced Ratings and Faculty Inter-Rater Reliability |
title | Is It All About the Form? Norm- vs Criterion-Referenced Ratings and Faculty Inter-Rater Reliability |
title_full | Is It All About the Form? Norm- vs Criterion-Referenced Ratings and Faculty Inter-Rater Reliability |
title_fullStr | Is It All About the Form? Norm- vs Criterion-Referenced Ratings and Faculty Inter-Rater Reliability |
title_full_unstemmed | Is It All About the Form? Norm- vs Criterion-Referenced Ratings and Faculty Inter-Rater Reliability |
title_short | Is It All About the Form? Norm- vs Criterion-Referenced Ratings and Faculty Inter-Rater Reliability |
title_sort | is it all about the form? norm- vs criterion-referenced ratings and faculty inter-rater reliability |
topic | Original Research |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10498947/ https://www.ncbi.nlm.nih.gov/pubmed/37711480 http://dx.doi.org/10.31486/toj.23.0014 |
work_keys_str_mv | AT scielzoshannona isitallabouttheformnormvscriterionreferencedratingsandfacultyinterraterreliability AT abdelfattahkareem isitallabouttheformnormvscriterionreferencedratingsandfacultyinterraterreliability AT ryderhilaryf isitallabouttheformnormvscriterionreferencedratingsandfacultyinterraterreliability |