Cargando…
Imitation learning for improved 3D PET/MR attenuation correction()
The assessment of the quality of synthesised/pseudo Computed Tomography (pCT) images is commonly measured by an intensity-wise similarity between the ground truth CT and the pCT. However, when using the pCT as an attenuation map ([Formula: see text]-map) for PET reconstruction in Positron Emission T...
Autores principales: | , , , , , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Elsevier
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7611431/ https://www.ncbi.nlm.nih.gov/pubmed/33951598 http://dx.doi.org/10.1016/j.media.2021.102079 |
_version_ | 1783605276254928896 |
---|---|
author | Kläser, Kerstin Varsavsky, Thomas Markiewicz, Pawel Vercauteren, Tom Hammers, Alexander Atkinson, David Thielemans, Kris Hutton, Brian Cardoso, M.J. Ourselin, Sébastien |
author_facet | Kläser, Kerstin Varsavsky, Thomas Markiewicz, Pawel Vercauteren, Tom Hammers, Alexander Atkinson, David Thielemans, Kris Hutton, Brian Cardoso, M.J. Ourselin, Sébastien |
author_sort | Kläser, Kerstin |
collection | PubMed |
description | The assessment of the quality of synthesised/pseudo Computed Tomography (pCT) images is commonly measured by an intensity-wise similarity between the ground truth CT and the pCT. However, when using the pCT as an attenuation map ([Formula: see text]-map) for PET reconstruction in Positron Emission Tomography Magnetic Resonance Imaging (PET/MRI) minimising the error between pCT and CT neglects the main objective of predicting a pCT that when used as [Formula: see text]-map reconstructs a pseudo PET (pPET) which is as similar as possible to the gold standard CT-derived PET reconstruction. This observation motivated us to propose a novel multi-hypothesis deep learning framework explicitly aimed at PET reconstruction application. A convolutional neural network (CNN) synthesises pCTs by minimising a combination of the pixel-wise error between pCT and CT and a novel metric-loss that itself is defined by a CNN and aims to minimise consequent PET residuals. Training is performed on a database of twenty 3D MR/CT/PET brain image pairs. Quantitative results on a fully independent dataset of twenty-three 3D MR/CT/PET image pairs show that the network is able to synthesise more accurate pCTs. The Mean Absolute Error on the pCT (110.98 HU [Formula: see text] 19.22 HU) compared to a baseline CNN (172.12 HU [Formula: see text] 19.61 HU) and a multi-atlas propagation approach (153.40 HU [Formula: see text] 18.68 HU), and subsequently lead to a significant improvement in the PET reconstruction error (4.74% [Formula: see text] 1.52% compared to baseline 13.72% [Formula: see text] 2.48% and multi-atlas propagation 6.68% [Formula: see text] 2.06%). |
format | Online Article Text |
id | pubmed-7611431 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | Elsevier |
record_format | MEDLINE/PubMed |
spelling | pubmed-76114312021-10-29 Imitation learning for improved 3D PET/MR attenuation correction() Kläser, Kerstin Varsavsky, Thomas Markiewicz, Pawel Vercauteren, Tom Hammers, Alexander Atkinson, David Thielemans, Kris Hutton, Brian Cardoso, M.J. Ourselin, Sébastien Med Image Anal Article The assessment of the quality of synthesised/pseudo Computed Tomography (pCT) images is commonly measured by an intensity-wise similarity between the ground truth CT and the pCT. However, when using the pCT as an attenuation map ([Formula: see text]-map) for PET reconstruction in Positron Emission Tomography Magnetic Resonance Imaging (PET/MRI) minimising the error between pCT and CT neglects the main objective of predicting a pCT that when used as [Formula: see text]-map reconstructs a pseudo PET (pPET) which is as similar as possible to the gold standard CT-derived PET reconstruction. This observation motivated us to propose a novel multi-hypothesis deep learning framework explicitly aimed at PET reconstruction application. A convolutional neural network (CNN) synthesises pCTs by minimising a combination of the pixel-wise error between pCT and CT and a novel metric-loss that itself is defined by a CNN and aims to minimise consequent PET residuals. Training is performed on a database of twenty 3D MR/CT/PET brain image pairs. Quantitative results on a fully independent dataset of twenty-three 3D MR/CT/PET image pairs show that the network is able to synthesise more accurate pCTs. The Mean Absolute Error on the pCT (110.98 HU [Formula: see text] 19.22 HU) compared to a baseline CNN (172.12 HU [Formula: see text] 19.61 HU) and a multi-atlas propagation approach (153.40 HU [Formula: see text] 18.68 HU), and subsequently lead to a significant improvement in the PET reconstruction error (4.74% [Formula: see text] 1.52% compared to baseline 13.72% [Formula: see text] 2.48% and multi-atlas propagation 6.68% [Formula: see text] 2.06%). Elsevier 2021-07 /pmc/articles/PMC7611431/ /pubmed/33951598 http://dx.doi.org/10.1016/j.media.2021.102079 Text en © 2021 The Author(s) https://creativecommons.org/licenses/by/4.0/This is an open access article under the CC BY license (http://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Kläser, Kerstin Varsavsky, Thomas Markiewicz, Pawel Vercauteren, Tom Hammers, Alexander Atkinson, David Thielemans, Kris Hutton, Brian Cardoso, M.J. Ourselin, Sébastien Imitation learning for improved 3D PET/MR attenuation correction() |
title | Imitation learning for improved 3D PET/MR attenuation correction() |
title_full | Imitation learning for improved 3D PET/MR attenuation correction() |
title_fullStr | Imitation learning for improved 3D PET/MR attenuation correction() |
title_full_unstemmed | Imitation learning for improved 3D PET/MR attenuation correction() |
title_short | Imitation learning for improved 3D PET/MR attenuation correction() |
title_sort | imitation learning for improved 3d pet/mr attenuation correction() |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7611431/ https://www.ncbi.nlm.nih.gov/pubmed/33951598 http://dx.doi.org/10.1016/j.media.2021.102079 |
work_keys_str_mv | AT klaserkerstin imitationlearningforimproved3dpetmrattenuationcorrection AT varsavskythomas imitationlearningforimproved3dpetmrattenuationcorrection AT markiewiczpawel imitationlearningforimproved3dpetmrattenuationcorrection AT vercauterentom imitationlearningforimproved3dpetmrattenuationcorrection AT hammersalexander imitationlearningforimproved3dpetmrattenuationcorrection AT atkinsondavid imitationlearningforimproved3dpetmrattenuationcorrection AT thielemanskris imitationlearningforimproved3dpetmrattenuationcorrection AT huttonbrian imitationlearningforimproved3dpetmrattenuationcorrection AT cardosomj imitationlearningforimproved3dpetmrattenuationcorrection AT ourselinsebastien imitationlearningforimproved3dpetmrattenuationcorrection |