Cargando…

Learning a preconditioner to accelerate compressed sensing reconstructions in MRI

PURPOSE: To learn a preconditioner that accelerates parallel imaging (PI) and compressed sensing (CS) reconstructions. METHODS: A convolutional neural network (CNN) with residual connections was used to train a preconditioning operator. Training and validation data were simulated using 50% brain ima...

Descripción completa

Detalles Bibliográficos
Autores principales: Koolstra, Kirsten, Remis, Rob
Formato: Online Artículo Texto
Lenguaje:English
Publicado: John Wiley and Sons Inc. 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9299023/
https://www.ncbi.nlm.nih.gov/pubmed/34752655
http://dx.doi.org/10.1002/mrm.29073
_version_ 1784750849738670080
author Koolstra, Kirsten
Remis, Rob
author_facet Koolstra, Kirsten
Remis, Rob
author_sort Koolstra, Kirsten
collection PubMed
description PURPOSE: To learn a preconditioner that accelerates parallel imaging (PI) and compressed sensing (CS) reconstructions. METHODS: A convolutional neural network (CNN) with residual connections was used to train a preconditioning operator. Training and validation data were simulated using 50% brain images and 50% white Gaussian noise images. Each multichannel training example contains a simulated sampling mask, complex coil sensitivity maps, and two regularization parameter maps. The trained model was integrated in the preconditioned conjugate gradient (PCG) method as part of the split Bregman CS method. The acceleration performance was compared with that of a circulant PI‐CS preconditioner for varying undersampling factors, number of coil elements and anatomies. RESULTS: The learned preconditioner reduces the number of PCG iterations by a factor of 4, yielding a similar acceleration as an efficient circulant preconditioner. The method generalizes well to different sampling schemes, coil configurations and anatomies. CONCLUSION: It is possible to learn adaptable preconditioners for PI and CS reconstructions that meet the performance of state‐of‐the‐art preconditioners. Further acceleration could be achieved by optimizing the network architecture and the training set. Such a preconditioner could also be integrated in fully learned reconstruction methods to accelerate the training process of unrolled networks.
format Online
Article
Text
id pubmed-9299023
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher John Wiley and Sons Inc.
record_format MEDLINE/PubMed
spelling pubmed-92990232022-07-21 Learning a preconditioner to accelerate compressed sensing reconstructions in MRI Koolstra, Kirsten Remis, Rob Magn Reson Med Technical Notes—Computer Processing and Modeling PURPOSE: To learn a preconditioner that accelerates parallel imaging (PI) and compressed sensing (CS) reconstructions. METHODS: A convolutional neural network (CNN) with residual connections was used to train a preconditioning operator. Training and validation data were simulated using 50% brain images and 50% white Gaussian noise images. Each multichannel training example contains a simulated sampling mask, complex coil sensitivity maps, and two regularization parameter maps. The trained model was integrated in the preconditioned conjugate gradient (PCG) method as part of the split Bregman CS method. The acceleration performance was compared with that of a circulant PI‐CS preconditioner for varying undersampling factors, number of coil elements and anatomies. RESULTS: The learned preconditioner reduces the number of PCG iterations by a factor of 4, yielding a similar acceleration as an efficient circulant preconditioner. The method generalizes well to different sampling schemes, coil configurations and anatomies. CONCLUSION: It is possible to learn adaptable preconditioners for PI and CS reconstructions that meet the performance of state‐of‐the‐art preconditioners. Further acceleration could be achieved by optimizing the network architecture and the training set. Such a preconditioner could also be integrated in fully learned reconstruction methods to accelerate the training process of unrolled networks. John Wiley and Sons Inc. 2021-11-09 2022-04 /pmc/articles/PMC9299023/ /pubmed/34752655 http://dx.doi.org/10.1002/mrm.29073 Text en © 2021 The Authors. Magnetic Resonance in Medicine published by Wiley Periodicals LLC on behalf of International Society for Magnetic Resonance in Medicine. https://creativecommons.org/licenses/by-nc/4.0/This is an open access article under the terms of the http://creativecommons.org/licenses/by-nc/4.0/ (https://creativecommons.org/licenses/by-nc/4.0/) License, which permits use, distribution and reproduction in any medium, provided the original work is properly cited and is not used for commercial purposes.
spellingShingle Technical Notes—Computer Processing and Modeling
Koolstra, Kirsten
Remis, Rob
Learning a preconditioner to accelerate compressed sensing reconstructions in MRI
title Learning a preconditioner to accelerate compressed sensing reconstructions in MRI
title_full Learning a preconditioner to accelerate compressed sensing reconstructions in MRI
title_fullStr Learning a preconditioner to accelerate compressed sensing reconstructions in MRI
title_full_unstemmed Learning a preconditioner to accelerate compressed sensing reconstructions in MRI
title_short Learning a preconditioner to accelerate compressed sensing reconstructions in MRI
title_sort learning a preconditioner to accelerate compressed sensing reconstructions in mri
topic Technical Notes—Computer Processing and Modeling
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9299023/
https://www.ncbi.nlm.nih.gov/pubmed/34752655
http://dx.doi.org/10.1002/mrm.29073
work_keys_str_mv AT koolstrakirsten learningapreconditionertoacceleratecompressedsensingreconstructionsinmri
AT remisrob learningapreconditionertoacceleratecompressedsensingreconstructionsinmri