Cargando…

A workstation-integrated peer review quality assurance program: pilot study

BACKGROUND: The surrogate indicator of radiological excellence that has become accepted is consistency of assessments between radiologists, and the technique that has become the standard for evaluating concordance is peer review. This study describes the results of a workstation-integrated peer revi...

Descripción completa

Detalles Bibliográficos
Autores principales: O’Keeffe, Margaret M, Davis, Todd M, Siminoski, Kerry
Formato: Online Artículo Texto
Lenguaje:English
Publicado: BioMed Central 2013
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3711932/
https://www.ncbi.nlm.nih.gov/pubmed/23822583
http://dx.doi.org/10.1186/1471-2342-13-19
_version_ 1782276992143458304
author O’Keeffe, Margaret M
Davis, Todd M
Siminoski, Kerry
author_facet O’Keeffe, Margaret M
Davis, Todd M
Siminoski, Kerry
author_sort O’Keeffe, Margaret M
collection PubMed
description BACKGROUND: The surrogate indicator of radiological excellence that has become accepted is consistency of assessments between radiologists, and the technique that has become the standard for evaluating concordance is peer review. This study describes the results of a workstation-integrated peer review program in a busy outpatient radiology practice. METHODS: Workstation-based peer review was performed using the software program Intelerad Peer Review. Cases for review were randomly chosen from those being actively reported. If an appropriate prior study was available, and if the reviewing radiologist and the original interpreting radiologist had not exceeded review targets, the case was scored using the modified RADPEER system. RESULTS: There were 2,241 cases randomly assigned for peer review. Of selected cases, 1,705 (76%) were interpreted. Reviewing radiologists agreed with prior reports in 99.1% of assessments. Positive feedback (score 0) was given in three cases (0.2%) and concordance (scores of 0 to 2) was assigned in 99.4%, similar to reported rates of 97.0% to 99.8%. Clinically significant discrepancies (scores of 3 or 4) were identified in 10 cases (0.6%). Eighty-eight percent of reviewed radiologists found the reviews worthwhile, 79% found scores appropriate, and 65% felt feedback was appropriate. Two-thirds of radiologists found case rounds discussing significant discrepancies to be valuable. CONCLUSIONS: The workstation-based computerized peer review process used in this pilot project was seamlessly incorporated into the normal workday and met most criteria for an ideal peer review system. Clinically significant discrepancies were identified in 0.6% of cases, similar to published outcomes using the RADPEER system. Reviewed radiologists felt the process was worthwhile.
format Online
Article
Text
id pubmed-3711932
institution National Center for Biotechnology Information
language English
publishDate 2013
publisher BioMed Central
record_format MEDLINE/PubMed
spelling pubmed-37119322013-07-16 A workstation-integrated peer review quality assurance program: pilot study O’Keeffe, Margaret M Davis, Todd M Siminoski, Kerry BMC Med Imaging Research Article BACKGROUND: The surrogate indicator of radiological excellence that has become accepted is consistency of assessments between radiologists, and the technique that has become the standard for evaluating concordance is peer review. This study describes the results of a workstation-integrated peer review program in a busy outpatient radiology practice. METHODS: Workstation-based peer review was performed using the software program Intelerad Peer Review. Cases for review were randomly chosen from those being actively reported. If an appropriate prior study was available, and if the reviewing radiologist and the original interpreting radiologist had not exceeded review targets, the case was scored using the modified RADPEER system. RESULTS: There were 2,241 cases randomly assigned for peer review. Of selected cases, 1,705 (76%) were interpreted. Reviewing radiologists agreed with prior reports in 99.1% of assessments. Positive feedback (score 0) was given in three cases (0.2%) and concordance (scores of 0 to 2) was assigned in 99.4%, similar to reported rates of 97.0% to 99.8%. Clinically significant discrepancies (scores of 3 or 4) were identified in 10 cases (0.6%). Eighty-eight percent of reviewed radiologists found the reviews worthwhile, 79% found scores appropriate, and 65% felt feedback was appropriate. Two-thirds of radiologists found case rounds discussing significant discrepancies to be valuable. CONCLUSIONS: The workstation-based computerized peer review process used in this pilot project was seamlessly incorporated into the normal workday and met most criteria for an ideal peer review system. Clinically significant discrepancies were identified in 0.6% of cases, similar to published outcomes using the RADPEER system. Reviewed radiologists felt the process was worthwhile. BioMed Central 2013-07-04 /pmc/articles/PMC3711932/ /pubmed/23822583 http://dx.doi.org/10.1186/1471-2342-13-19 Text en Copyright © 2013 O’Keeffe et al.; licensee BioMed Central Ltd. http://creativecommons.org/licenses/by/2.0 This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
spellingShingle Research Article
O’Keeffe, Margaret M
Davis, Todd M
Siminoski, Kerry
A workstation-integrated peer review quality assurance program: pilot study
title A workstation-integrated peer review quality assurance program: pilot study
title_full A workstation-integrated peer review quality assurance program: pilot study
title_fullStr A workstation-integrated peer review quality assurance program: pilot study
title_full_unstemmed A workstation-integrated peer review quality assurance program: pilot study
title_short A workstation-integrated peer review quality assurance program: pilot study
title_sort workstation-integrated peer review quality assurance program: pilot study
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3711932/
https://www.ncbi.nlm.nih.gov/pubmed/23822583
http://dx.doi.org/10.1186/1471-2342-13-19
work_keys_str_mv AT okeeffemargaretm aworkstationintegratedpeerreviewqualityassuranceprogrampilotstudy
AT davistoddm aworkstationintegratedpeerreviewqualityassuranceprogrampilotstudy
AT siminoskikerry aworkstationintegratedpeerreviewqualityassuranceprogrampilotstudy
AT okeeffemargaretm workstationintegratedpeerreviewqualityassuranceprogrampilotstudy
AT davistoddm workstationintegratedpeerreviewqualityassuranceprogrampilotstudy
AT siminoskikerry workstationintegratedpeerreviewqualityassuranceprogrampilotstudy