Cargando…

Neural network-assisted automated image registration for MRI-guided adaptive brachytherapy in cervical cancer

PURPOSE: In image-guided adaptive brachytherapy (IGABT) a quantitative evaluation of the dosimetric changes between fractions due to anatomical variations, can be implemented via rigid registration of images from subsequent fractions based on the applicator as a reference structure. With available t...

Descripción completa

Detalles Bibliográficos
Autores principales: Ecker, Stefan, Zimmermann, Lukas, Heilemann, Gerd, Niatsetski, Yury, Schmid, Maximilian, Sturdza, Alina Emiliana, Knoth, Johannes, Kirisits, Christian, Nesvacil, Nicole
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Elsevier 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9948828/
https://www.ncbi.nlm.nih.gov/pubmed/35570099
http://dx.doi.org/10.1016/j.zemedi.2022.04.002
Descripción
Sumario:PURPOSE: In image-guided adaptive brachytherapy (IGABT) a quantitative evaluation of the dosimetric changes between fractions due to anatomical variations, can be implemented via rigid registration of images from subsequent fractions based on the applicator as a reference structure. With available treatment planning systems (TPS), this is a manual and time-consuming process. The aim of this retrospective study was to automate this process. A neural network (NN) was trained to predict the applicator structure from MR images. The resulting segmentation was used to automatically register MR-volumes. MATERIAL AND METHODS: DICOM images and plans of 56 patients treated for cervical cancer with high dose-rate (HDR) brachytherapy were used in the study. A 2D and a 3D NN were trained to segment applicator structures on clinical T2-weighted MRI datasets. Different rigid registration algorithms were investigated and compared. To evaluate a fully automatic registration workflow, the NN-predicted applicator segmentations (AS) were used for rigid image registration with the best performing algorithm. The DICE coefficient and mean distance error between dwell positions (MDE) were used to evaluate segmentation and registration performance. RESULTS: The mean DICE coefficient for the predicted AS was 0.70 ± 0.07 and 0.58 ± 0.04 for the 3D NN and 2D NN, respectively. Registration algorithms achieved MDE errors from 8.1 ± 3.7 mm (worst) to 0.7 ± 0.5 mm (best), using ground-truth AS. Using the predicted AS from the 3D NN together with the best registration algorithm, an MDE of 2.7 ± 1.4 mm was achieved. CONCLUSION: Using a combination of deep learning models and state of the art image registration techniques has been demonstrated to be a promising solution for automatic image registration in IGABT. In combination with auto-contouring of organs at risk, the auto-registration workflow from this study could become part of an online-dosimetric interfraction evaluation workflow in the future.