Cargando…

Deep Learning-Assisted Diagnosis of Pediatric Skull Fractures on Plain Radiographs

OBJECTIVE: To develop and evaluate a deep learning-based artificial intelligence (AI) model for detecting skull fractures on plain radiographs in children. MATERIALS AND METHODS: This retrospective multi-center study consisted of a development dataset acquired from two hospitals (n = 149 and 264) an...

Descripción completa

Detalles Bibliográficos
Autores principales: Choi, Jae Won, Cho, Yeon Jin, Ha, Ji Young, Lee, Yun Young, Koh, Seok Young, Seo, June Young, Choi, Young Hun, Cheon, Jung-Eun, Phi, Ji Hoon, Kim, Injoon, Yang, Jaekwang, Kim, Woo Sun
Formato: Online Artículo Texto
Lenguaje:English
Publicado: The Korean Society of Radiology 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8876653/
https://www.ncbi.nlm.nih.gov/pubmed/35029078
http://dx.doi.org/10.3348/kjr.2021.0449
Descripción
Sumario:OBJECTIVE: To develop and evaluate a deep learning-based artificial intelligence (AI) model for detecting skull fractures on plain radiographs in children. MATERIALS AND METHODS: This retrospective multi-center study consisted of a development dataset acquired from two hospitals (n = 149 and 264) and an external test set (n = 95) from a third hospital. Datasets included children with head trauma who underwent both skull radiography and cranial computed tomography (CT). The development dataset was split into training, tuning, and internal test sets in a ratio of 7:1:2. The reference standard for skull fracture was cranial CT. Two radiology residents, a pediatric radiologist, and two emergency physicians participated in a two-session observer study on an external test set with and without AI assistance. We obtained the area under the receiver operating characteristic curve (AUROC), sensitivity, and specificity along with their 95% confidence intervals (CIs). RESULTS: The AI model showed an AUROC of 0.922 (95% CI, 0.842–0.969) in the internal test set and 0.870 (95% CI, 0.785–0.930) in the external test set. The model had a sensitivity of 81.1% (95% CI, 64.8%–92.0%) and specificity of 91.3% (95% CI, 79.2%–97.6%) for the internal test set and 78.9% (95% CI, 54.4%–93.9%) and 88.2% (95% CI, 78.7%–94.4%), respectively, for the external test set. With the model’s assistance, significant AUROC improvement was observed in radiology residents (pooled results) and emergency physicians (pooled results) with the difference from reading without AI assistance of 0.094 (95% CI, 0.020–0.168; p = 0.012) and 0.069 (95% CI, 0.002–0.136; p = 0.043), respectively, but not in the pediatric radiologist with the difference of 0.008 (95% CI, -0.074–0.090; p = 0.850). CONCLUSION: A deep learning-based AI model improved the performance of inexperienced radiologists and emergency physicians in diagnosing pediatric skull fractures on plain radiographs.