Cargando…

Can we predict discordant RECIST 1.1 evaluations in double read clinical trials?

BACKGROUND: In lung clinical trials with imaging, blinded independent central review with double reads is recommended to reduce evaluation bias and the Response Evaluation Criteria In Solid Tumor (RECIST) is still widely used. We retrospectively analyzed the inter-reader discrepancies rate over time...

Descripción completa

Detalles Bibliográficos
Autores principales: Beaumont, Hubert, Iannessi, Antoine
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10585359/
https://www.ncbi.nlm.nih.gov/pubmed/37869080
http://dx.doi.org/10.3389/fonc.2023.1239570
_version_ 1785122938568048640
author Beaumont, Hubert
Iannessi, Antoine
author_facet Beaumont, Hubert
Iannessi, Antoine
author_sort Beaumont, Hubert
collection PubMed
description BACKGROUND: In lung clinical trials with imaging, blinded independent central review with double reads is recommended to reduce evaluation bias and the Response Evaluation Criteria In Solid Tumor (RECIST) is still widely used. We retrospectively analyzed the inter-reader discrepancies rate over time, the risk factors for discrepancies related to baseline evaluations, and the potential of machine learning to predict inter-reader discrepancies. MATERIALS AND METHODS: We retrospectively analyzed five BICR clinical trials for patients on immunotherapy or targeted therapy for lung cancer. Double reads of 1724 patients involving 17 radiologists were performed using RECIST 1.1. We evaluated the rate of discrepancies over time according to four endpoints: progressive disease declared (PDD), date of progressive disease (DOPD), best overall response (BOR), and date of the first response (DOFR). Risk factors associated with discrepancies were analyzed, two predictive models were evaluated. RESULTS: At the end of trials, the discrepancy rates between trials were not different. On average, the discrepancy rates were 21.0%, 41.0%, 28.8%, and 48.8% for PDD, DOPD, BOR, and DOFR, respectively. Over time, the discrepancy rate was higher for DOFR than DOPD, and the rates increased as the trial progressed, even after accrual was completed. It was rare for readers to not find any disease, for less than 7% of patients, at least one reader selected non-measurable disease only (NTL). Often the readers selected some of their target lesions (TLs) and NTLs in different organs, with ranges of 36.0-57.9% and 60.5-73.5% of patients, respectively. Rarely (4-8.1%) two readers selected all their TLs in different locations. Significant risk factors were different depending on the endpoint and the trial being considered. Prediction had a poor performance but the positive predictive value was higher than 80%. The best classification was obtained with BOR. CONCLUSION: Predicting discordance rates necessitates having knowledge of patient accrual, patient survival, and the probability of discordances over time. In lung cancer trials, although risk factors for inter-reader discrepancies are known, they are weakly significant, the ability to predict discrepancies from baseline data is limited. To boost prediction accuracy, it would be necessary to enhance baseline-derived features or create new ones, considering other risk factors and looking into optimal reader associations.
format Online
Article
Text
id pubmed-10585359
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-105853592023-10-20 Can we predict discordant RECIST 1.1 evaluations in double read clinical trials? Beaumont, Hubert Iannessi, Antoine Front Oncol Oncology BACKGROUND: In lung clinical trials with imaging, blinded independent central review with double reads is recommended to reduce evaluation bias and the Response Evaluation Criteria In Solid Tumor (RECIST) is still widely used. We retrospectively analyzed the inter-reader discrepancies rate over time, the risk factors for discrepancies related to baseline evaluations, and the potential of machine learning to predict inter-reader discrepancies. MATERIALS AND METHODS: We retrospectively analyzed five BICR clinical trials for patients on immunotherapy or targeted therapy for lung cancer. Double reads of 1724 patients involving 17 radiologists were performed using RECIST 1.1. We evaluated the rate of discrepancies over time according to four endpoints: progressive disease declared (PDD), date of progressive disease (DOPD), best overall response (BOR), and date of the first response (DOFR). Risk factors associated with discrepancies were analyzed, two predictive models were evaluated. RESULTS: At the end of trials, the discrepancy rates between trials were not different. On average, the discrepancy rates were 21.0%, 41.0%, 28.8%, and 48.8% for PDD, DOPD, BOR, and DOFR, respectively. Over time, the discrepancy rate was higher for DOFR than DOPD, and the rates increased as the trial progressed, even after accrual was completed. It was rare for readers to not find any disease, for less than 7% of patients, at least one reader selected non-measurable disease only (NTL). Often the readers selected some of their target lesions (TLs) and NTLs in different organs, with ranges of 36.0-57.9% and 60.5-73.5% of patients, respectively. Rarely (4-8.1%) two readers selected all their TLs in different locations. Significant risk factors were different depending on the endpoint and the trial being considered. Prediction had a poor performance but the positive predictive value was higher than 80%. The best classification was obtained with BOR. CONCLUSION: Predicting discordance rates necessitates having knowledge of patient accrual, patient survival, and the probability of discordances over time. In lung cancer trials, although risk factors for inter-reader discrepancies are known, they are weakly significant, the ability to predict discrepancies from baseline data is limited. To boost prediction accuracy, it would be necessary to enhance baseline-derived features or create new ones, considering other risk factors and looking into optimal reader associations. Frontiers Media S.A. 2023-10-04 /pmc/articles/PMC10585359/ /pubmed/37869080 http://dx.doi.org/10.3389/fonc.2023.1239570 Text en Copyright © 2023 Beaumont and Iannessi https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Oncology
Beaumont, Hubert
Iannessi, Antoine
Can we predict discordant RECIST 1.1 evaluations in double read clinical trials?
title Can we predict discordant RECIST 1.1 evaluations in double read clinical trials?
title_full Can we predict discordant RECIST 1.1 evaluations in double read clinical trials?
title_fullStr Can we predict discordant RECIST 1.1 evaluations in double read clinical trials?
title_full_unstemmed Can we predict discordant RECIST 1.1 evaluations in double read clinical trials?
title_short Can we predict discordant RECIST 1.1 evaluations in double read clinical trials?
title_sort can we predict discordant recist 1.1 evaluations in double read clinical trials?
topic Oncology
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10585359/
https://www.ncbi.nlm.nih.gov/pubmed/37869080
http://dx.doi.org/10.3389/fonc.2023.1239570
work_keys_str_mv AT beaumonthubert canwepredictdiscordantrecist11evaluationsindoublereadclinicaltrials
AT iannessiantoine canwepredictdiscordantrecist11evaluationsindoublereadclinicaltrials