Cargando…
E pluribus unum: prospective acceptability benchmarking from the Contouring Collaborative for Consensus in Radiation Oncology crowdsourced initiative for multiobserver segmentation
PURPOSE: Contouring Collaborative for Consensus in Radiation Oncology (C3RO) is a crowdsourced challenge engaging radiation oncologists across various expertise levels in segmentation. An obstacle to artificial intelligence (AI) development is the paucity of multiexpert datasets; consequently, we so...
Autores principales: | , , , , , , , , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Society of Photo-Optical Instrumentation Engineers
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9907021/ https://www.ncbi.nlm.nih.gov/pubmed/36761036 http://dx.doi.org/10.1117/1.JMI.10.S1.S11903 |
_version_ | 1784884086731440128 |
---|---|
author | Lin, Diana Wahid, Kareem A. Nelms, Benjamin E. He, Renjie Naser, Mohammed A. Duke, Simon Sherer, Michael V. Christodouleas, John P. Mohamed, Abdallah S. R. Cislo, Michael Murphy, James D. Fuller, Clifton D. Gillespie, Erin F. |
author_facet | Lin, Diana Wahid, Kareem A. Nelms, Benjamin E. He, Renjie Naser, Mohammed A. Duke, Simon Sherer, Michael V. Christodouleas, John P. Mohamed, Abdallah S. R. Cislo, Michael Murphy, James D. Fuller, Clifton D. Gillespie, Erin F. |
author_sort | Lin, Diana |
collection | PubMed |
description | PURPOSE: Contouring Collaborative for Consensus in Radiation Oncology (C3RO) is a crowdsourced challenge engaging radiation oncologists across various expertise levels in segmentation. An obstacle to artificial intelligence (AI) development is the paucity of multiexpert datasets; consequently, we sought to characterize whether aggregate segmentations generated from multiple nonexperts could meet or exceed recognized expert agreement. APPROACH: Participants who contoured [Formula: see text] region of interest (ROI) for the breast, sarcoma, head and neck (H&N), gynecologic (GYN), or gastrointestinal (GI) cases were identified as a nonexpert or recognized expert. Cohort-specific ROIs were combined into single simultaneous truth and performance level estimation (STAPLE) consensus segmentations. [Formula: see text] ROIs were evaluated against [Formula: see text] contours using Dice similarity coefficient (DSC). The expert interobserver DSC ([Formula: see text]) was calculated as an acceptability threshold between [Formula: see text] and [Formula: see text]. To determine the number of nonexperts required to match the [Formula: see text] for each ROI, a single consensus contour was generated using variable numbers of nonexperts and then compared to the [Formula: see text]. RESULTS: For all cases, the DSC values for [Formula: see text] versus [Formula: see text] were higher than comparator expert [Formula: see text] for most ROIs. The minimum number of nonexpert segmentations needed for a consensus ROI to achieve [Formula: see text] acceptability criteria ranged between 2 and 4 for breast, 3 and 5 for sarcoma, 3 and 5 for H&N, 3 and 5 for GYN, and 3 for GI. CONCLUSIONS: Multiple nonexpert-generated consensus ROIs met or exceeded expert-derived acceptability thresholds. Five nonexperts could potentially generate consensus segmentations for most ROIs with performance approximating experts, suggesting nonexpert segmentations as feasible cost-effective AI inputs. |
format | Online Article Text |
id | pubmed-9907021 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Society of Photo-Optical Instrumentation Engineers |
record_format | MEDLINE/PubMed |
spelling | pubmed-99070212023-02-08 E pluribus unum: prospective acceptability benchmarking from the Contouring Collaborative for Consensus in Radiation Oncology crowdsourced initiative for multiobserver segmentation Lin, Diana Wahid, Kareem A. Nelms, Benjamin E. He, Renjie Naser, Mohammed A. Duke, Simon Sherer, Michael V. Christodouleas, John P. Mohamed, Abdallah S. R. Cislo, Michael Murphy, James D. Fuller, Clifton D. Gillespie, Erin F. J Med Imaging (Bellingham) Special Issue on Medical Image Perception and Observer Performance PURPOSE: Contouring Collaborative for Consensus in Radiation Oncology (C3RO) is a crowdsourced challenge engaging radiation oncologists across various expertise levels in segmentation. An obstacle to artificial intelligence (AI) development is the paucity of multiexpert datasets; consequently, we sought to characterize whether aggregate segmentations generated from multiple nonexperts could meet or exceed recognized expert agreement. APPROACH: Participants who contoured [Formula: see text] region of interest (ROI) for the breast, sarcoma, head and neck (H&N), gynecologic (GYN), or gastrointestinal (GI) cases were identified as a nonexpert or recognized expert. Cohort-specific ROIs were combined into single simultaneous truth and performance level estimation (STAPLE) consensus segmentations. [Formula: see text] ROIs were evaluated against [Formula: see text] contours using Dice similarity coefficient (DSC). The expert interobserver DSC ([Formula: see text]) was calculated as an acceptability threshold between [Formula: see text] and [Formula: see text]. To determine the number of nonexperts required to match the [Formula: see text] for each ROI, a single consensus contour was generated using variable numbers of nonexperts and then compared to the [Formula: see text]. RESULTS: For all cases, the DSC values for [Formula: see text] versus [Formula: see text] were higher than comparator expert [Formula: see text] for most ROIs. The minimum number of nonexpert segmentations needed for a consensus ROI to achieve [Formula: see text] acceptability criteria ranged between 2 and 4 for breast, 3 and 5 for sarcoma, 3 and 5 for H&N, 3 and 5 for GYN, and 3 for GI. CONCLUSIONS: Multiple nonexpert-generated consensus ROIs met or exceeded expert-derived acceptability thresholds. Five nonexperts could potentially generate consensus segmentations for most ROIs with performance approximating experts, suggesting nonexpert segmentations as feasible cost-effective AI inputs. Society of Photo-Optical Instrumentation Engineers 2023-02-08 2023-02 /pmc/articles/PMC9907021/ /pubmed/36761036 http://dx.doi.org/10.1117/1.JMI.10.S1.S11903 Text en © 2023 The Authors https://creativecommons.org/licenses/by/4.0/Published by SPIE under a Creative Commons Attribution 4.0 International License. Distribution or reproduction of this work in whole or in part requires full attribution of the original publication, including its DOI. |
spellingShingle | Special Issue on Medical Image Perception and Observer Performance Lin, Diana Wahid, Kareem A. Nelms, Benjamin E. He, Renjie Naser, Mohammed A. Duke, Simon Sherer, Michael V. Christodouleas, John P. Mohamed, Abdallah S. R. Cislo, Michael Murphy, James D. Fuller, Clifton D. Gillespie, Erin F. E pluribus unum: prospective acceptability benchmarking from the Contouring Collaborative for Consensus in Radiation Oncology crowdsourced initiative for multiobserver segmentation |
title | E pluribus unum: prospective acceptability benchmarking from the Contouring Collaborative for Consensus in Radiation Oncology crowdsourced initiative for multiobserver segmentation |
title_full | E pluribus unum: prospective acceptability benchmarking from the Contouring Collaborative for Consensus in Radiation Oncology crowdsourced initiative for multiobserver segmentation |
title_fullStr | E pluribus unum: prospective acceptability benchmarking from the Contouring Collaborative for Consensus in Radiation Oncology crowdsourced initiative for multiobserver segmentation |
title_full_unstemmed | E pluribus unum: prospective acceptability benchmarking from the Contouring Collaborative for Consensus in Radiation Oncology crowdsourced initiative for multiobserver segmentation |
title_short | E pluribus unum: prospective acceptability benchmarking from the Contouring Collaborative for Consensus in Radiation Oncology crowdsourced initiative for multiobserver segmentation |
title_sort | e pluribus unum: prospective acceptability benchmarking from the contouring collaborative for consensus in radiation oncology crowdsourced initiative for multiobserver segmentation |
topic | Special Issue on Medical Image Perception and Observer Performance |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9907021/ https://www.ncbi.nlm.nih.gov/pubmed/36761036 http://dx.doi.org/10.1117/1.JMI.10.S1.S11903 |
work_keys_str_mv | AT lindiana epluribusunumprospectiveacceptabilitybenchmarkingfromthecontouringcollaborativeforconsensusinradiationoncologycrowdsourcedinitiativeformultiobserversegmentation AT wahidkareema epluribusunumprospectiveacceptabilitybenchmarkingfromthecontouringcollaborativeforconsensusinradiationoncologycrowdsourcedinitiativeformultiobserversegmentation AT nelmsbenjamine epluribusunumprospectiveacceptabilitybenchmarkingfromthecontouringcollaborativeforconsensusinradiationoncologycrowdsourcedinitiativeformultiobserversegmentation AT herenjie epluribusunumprospectiveacceptabilitybenchmarkingfromthecontouringcollaborativeforconsensusinradiationoncologycrowdsourcedinitiativeformultiobserversegmentation AT nasermohammeda epluribusunumprospectiveacceptabilitybenchmarkingfromthecontouringcollaborativeforconsensusinradiationoncologycrowdsourcedinitiativeformultiobserversegmentation AT dukesimon epluribusunumprospectiveacceptabilitybenchmarkingfromthecontouringcollaborativeforconsensusinradiationoncologycrowdsourcedinitiativeformultiobserversegmentation AT sherermichaelv epluribusunumprospectiveacceptabilitybenchmarkingfromthecontouringcollaborativeforconsensusinradiationoncologycrowdsourcedinitiativeformultiobserversegmentation AT christodouleasjohnp epluribusunumprospectiveacceptabilitybenchmarkingfromthecontouringcollaborativeforconsensusinradiationoncologycrowdsourcedinitiativeformultiobserversegmentation AT mohamedabdallahsr epluribusunumprospectiveacceptabilitybenchmarkingfromthecontouringcollaborativeforconsensusinradiationoncologycrowdsourcedinitiativeformultiobserversegmentation AT cislomichael epluribusunumprospectiveacceptabilitybenchmarkingfromthecontouringcollaborativeforconsensusinradiationoncologycrowdsourcedinitiativeformultiobserversegmentation AT murphyjamesd epluribusunumprospectiveacceptabilitybenchmarkingfromthecontouringcollaborativeforconsensusinradiationoncologycrowdsourcedinitiativeformultiobserversegmentation AT fullercliftond epluribusunumprospectiveacceptabilitybenchmarkingfromthecontouringcollaborativeforconsensusinradiationoncologycrowdsourcedinitiativeformultiobserversegmentation AT gillespieerinf epluribusunumprospectiveacceptabilitybenchmarkingfromthecontouringcollaborativeforconsensusinradiationoncologycrowdsourcedinitiativeformultiobserversegmentation |