Cargando…

Comparing 3D, 2.5D, and 2D Approaches to Brain Image Auto-Segmentation

Deep-learning methods for auto-segmenting brain images either segment one slice of the image (2D), five consecutive slices of the image (2.5D), or an entire volume of the image (3D). Whether one approach is superior for auto-segmenting brain images is not known. We compared these three approaches (3...

Descripción completa

Detalles Bibliográficos
Autores principales: Avesta, Arman, Hossain, Sajid, Lin, MingDe, Aboian, Mariam, Krumholz, Harlan M., Aneja, Sanjay
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9952534/
https://www.ncbi.nlm.nih.gov/pubmed/36829675
http://dx.doi.org/10.3390/bioengineering10020181
_version_ 1784893655198203904
author Avesta, Arman
Hossain, Sajid
Lin, MingDe
Aboian, Mariam
Krumholz, Harlan M.
Aneja, Sanjay
author_facet Avesta, Arman
Hossain, Sajid
Lin, MingDe
Aboian, Mariam
Krumholz, Harlan M.
Aneja, Sanjay
author_sort Avesta, Arman
collection PubMed
description Deep-learning methods for auto-segmenting brain images either segment one slice of the image (2D), five consecutive slices of the image (2.5D), or an entire volume of the image (3D). Whether one approach is superior for auto-segmenting brain images is not known. We compared these three approaches (3D, 2.5D, and 2D) across three auto-segmentation models (capsule networks, UNets, and nnUNets) to segment brain structures. We used 3430 brain MRIs, acquired in a multi-institutional study, to train and test our models. We used the following performance metrics: segmentation accuracy, performance with limited training data, required computational memory, and computational speed during training and deployment. The 3D, 2.5D, and 2D approaches respectively gave the highest to lowest Dice scores across all models. 3D models maintained higher Dice scores when the training set size was decreased from 3199 MRIs down to 60 MRIs. 3D models converged 20% to 40% faster during training and were 30% to 50% faster during deployment. However, 3D models require 20 times more computational memory compared to 2.5D or 2D models. This study showed that 3D models are more accurate, maintain better performance with limited training data, and are faster to train and deploy. However, 3D models require more computational memory compared to 2.5D or 2D models.
format Online
Article
Text
id pubmed-9952534
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-99525342023-02-25 Comparing 3D, 2.5D, and 2D Approaches to Brain Image Auto-Segmentation Avesta, Arman Hossain, Sajid Lin, MingDe Aboian, Mariam Krumholz, Harlan M. Aneja, Sanjay Bioengineering (Basel) Article Deep-learning methods for auto-segmenting brain images either segment one slice of the image (2D), five consecutive slices of the image (2.5D), or an entire volume of the image (3D). Whether one approach is superior for auto-segmenting brain images is not known. We compared these three approaches (3D, 2.5D, and 2D) across three auto-segmentation models (capsule networks, UNets, and nnUNets) to segment brain structures. We used 3430 brain MRIs, acquired in a multi-institutional study, to train and test our models. We used the following performance metrics: segmentation accuracy, performance with limited training data, required computational memory, and computational speed during training and deployment. The 3D, 2.5D, and 2D approaches respectively gave the highest to lowest Dice scores across all models. 3D models maintained higher Dice scores when the training set size was decreased from 3199 MRIs down to 60 MRIs. 3D models converged 20% to 40% faster during training and were 30% to 50% faster during deployment. However, 3D models require 20 times more computational memory compared to 2.5D or 2D models. This study showed that 3D models are more accurate, maintain better performance with limited training data, and are faster to train and deploy. However, 3D models require more computational memory compared to 2.5D or 2D models. MDPI 2023-02-01 /pmc/articles/PMC9952534/ /pubmed/36829675 http://dx.doi.org/10.3390/bioengineering10020181 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Avesta, Arman
Hossain, Sajid
Lin, MingDe
Aboian, Mariam
Krumholz, Harlan M.
Aneja, Sanjay
Comparing 3D, 2.5D, and 2D Approaches to Brain Image Auto-Segmentation
title Comparing 3D, 2.5D, and 2D Approaches to Brain Image Auto-Segmentation
title_full Comparing 3D, 2.5D, and 2D Approaches to Brain Image Auto-Segmentation
title_fullStr Comparing 3D, 2.5D, and 2D Approaches to Brain Image Auto-Segmentation
title_full_unstemmed Comparing 3D, 2.5D, and 2D Approaches to Brain Image Auto-Segmentation
title_short Comparing 3D, 2.5D, and 2D Approaches to Brain Image Auto-Segmentation
title_sort comparing 3d, 2.5d, and 2d approaches to brain image auto-segmentation
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9952534/
https://www.ncbi.nlm.nih.gov/pubmed/36829675
http://dx.doi.org/10.3390/bioengineering10020181
work_keys_str_mv AT avestaarman comparing3d25dand2dapproachestobrainimageautosegmentation
AT hossainsajid comparing3d25dand2dapproachestobrainimageautosegmentation
AT linmingde comparing3d25dand2dapproachestobrainimageautosegmentation
AT aboianmariam comparing3d25dand2dapproachestobrainimageautosegmentation
AT krumholzharlanm comparing3d25dand2dapproachestobrainimageautosegmentation
AT anejasanjay comparing3d25dand2dapproachestobrainimageautosegmentation