Cargando…

MAE-TransRNet: An improved transformer-ConvNet architecture with masked autoencoder for cardiac MRI registration

The heart is a relatively complex non-rigid motion organ in the human body. Quantitative motion analysis of the heart takes on a critical significance to help doctors with accurate diagnosis and treatment. Moreover, cardiovascular magnetic resonance imaging (CMRI) can be used to perform a more detai...

Descripción completa

Detalles Bibliográficos
Autores principales: Xiao, Xin, Dong, Suyu, Yu, Yang, Li, Yan, Yang, Guangyuan, Qiu, Zhaowen
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10033952/
https://www.ncbi.nlm.nih.gov/pubmed/36968818
http://dx.doi.org/10.3389/fmed.2023.1114571
_version_ 1784911100701048832
author Xiao, Xin
Dong, Suyu
Yu, Yang
Li, Yan
Yang, Guangyuan
Qiu, Zhaowen
author_facet Xiao, Xin
Dong, Suyu
Yu, Yang
Li, Yan
Yang, Guangyuan
Qiu, Zhaowen
author_sort Xiao, Xin
collection PubMed
description The heart is a relatively complex non-rigid motion organ in the human body. Quantitative motion analysis of the heart takes on a critical significance to help doctors with accurate diagnosis and treatment. Moreover, cardiovascular magnetic resonance imaging (CMRI) can be used to perform a more detailed quantitative analysis evaluation for cardiac diagnosis. Deformable image registration (DIR) has become a vital task in biomedical image analysis since tissue structures have variability in medical images. Recently, the model based on masked autoencoder (MAE) has recently been shown to be effective in computer vision tasks. Vision Transformer has the context aggregation ability to restore the semantic information in the original image regions by using a low proportion of visible image patches to predict the masked image patches. A novel Transformer-ConvNet architecture is proposed in this study based on MAE for medical image registration. The core of the Transformer is designed as a masked autoencoder (MAE) and a lightweight decoder structure, and feature extraction before the downstream registration task is transformed into the self-supervised learning task. This study also rethinks the calculation method of the multi-head self-attention mechanism in the Transformer encoder. We improve the query-key-value-based dot product attention by introducing both depthwise separable convolution (DWSC) and squeeze and excitation (SE) modules into the self-attention module to reduce the amount of parameter computation to highlight image details and maintain high spatial resolution image features. In addition, concurrent spatial and channel squeeze and excitation (scSE) module is embedded into the CNN structure, which also proves to be effective for extracting robust feature representations. The proposed method, called MAE-TransRNet, has better generalization. The proposed model is evaluated on the cardiac short-axis public dataset (with images and labels) at the 2017 Automated Cardiac Diagnosis Challenge (ACDC). The relevant qualitative and quantitative results (e.g., dice performance and Hausdorff distance) suggest that the proposed model can achieve superior results over those achieved by the state-of-the-art methods, thus proving that MAE and improved self-attention are more effective and promising for medical image registration tasks. Codes and models are available at https://github.com/XinXiao101/MAE-TransRNet.
format Online
Article
Text
id pubmed-10033952
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-100339522023-03-24 MAE-TransRNet: An improved transformer-ConvNet architecture with masked autoencoder for cardiac MRI registration Xiao, Xin Dong, Suyu Yu, Yang Li, Yan Yang, Guangyuan Qiu, Zhaowen Front Med (Lausanne) Medicine The heart is a relatively complex non-rigid motion organ in the human body. Quantitative motion analysis of the heart takes on a critical significance to help doctors with accurate diagnosis and treatment. Moreover, cardiovascular magnetic resonance imaging (CMRI) can be used to perform a more detailed quantitative analysis evaluation for cardiac diagnosis. Deformable image registration (DIR) has become a vital task in biomedical image analysis since tissue structures have variability in medical images. Recently, the model based on masked autoencoder (MAE) has recently been shown to be effective in computer vision tasks. Vision Transformer has the context aggregation ability to restore the semantic information in the original image regions by using a low proportion of visible image patches to predict the masked image patches. A novel Transformer-ConvNet architecture is proposed in this study based on MAE for medical image registration. The core of the Transformer is designed as a masked autoencoder (MAE) and a lightweight decoder structure, and feature extraction before the downstream registration task is transformed into the self-supervised learning task. This study also rethinks the calculation method of the multi-head self-attention mechanism in the Transformer encoder. We improve the query-key-value-based dot product attention by introducing both depthwise separable convolution (DWSC) and squeeze and excitation (SE) modules into the self-attention module to reduce the amount of parameter computation to highlight image details and maintain high spatial resolution image features. In addition, concurrent spatial and channel squeeze and excitation (scSE) module is embedded into the CNN structure, which also proves to be effective for extracting robust feature representations. The proposed method, called MAE-TransRNet, has better generalization. The proposed model is evaluated on the cardiac short-axis public dataset (with images and labels) at the 2017 Automated Cardiac Diagnosis Challenge (ACDC). The relevant qualitative and quantitative results (e.g., dice performance and Hausdorff distance) suggest that the proposed model can achieve superior results over those achieved by the state-of-the-art methods, thus proving that MAE and improved self-attention are more effective and promising for medical image registration tasks. Codes and models are available at https://github.com/XinXiao101/MAE-TransRNet. Frontiers Media S.A. 2023-03-09 /pmc/articles/PMC10033952/ /pubmed/36968818 http://dx.doi.org/10.3389/fmed.2023.1114571 Text en Copyright © 2023 Xiao, Dong, Yu, Li, Yang and Qiu. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Medicine
Xiao, Xin
Dong, Suyu
Yu, Yang
Li, Yan
Yang, Guangyuan
Qiu, Zhaowen
MAE-TransRNet: An improved transformer-ConvNet architecture with masked autoencoder for cardiac MRI registration
title MAE-TransRNet: An improved transformer-ConvNet architecture with masked autoencoder for cardiac MRI registration
title_full MAE-TransRNet: An improved transformer-ConvNet architecture with masked autoencoder for cardiac MRI registration
title_fullStr MAE-TransRNet: An improved transformer-ConvNet architecture with masked autoencoder for cardiac MRI registration
title_full_unstemmed MAE-TransRNet: An improved transformer-ConvNet architecture with masked autoencoder for cardiac MRI registration
title_short MAE-TransRNet: An improved transformer-ConvNet architecture with masked autoencoder for cardiac MRI registration
title_sort mae-transrnet: an improved transformer-convnet architecture with masked autoencoder for cardiac mri registration
topic Medicine
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10033952/
https://www.ncbi.nlm.nih.gov/pubmed/36968818
http://dx.doi.org/10.3389/fmed.2023.1114571
work_keys_str_mv AT xiaoxin maetransrnetanimprovedtransformerconvnetarchitecturewithmaskedautoencoderforcardiacmriregistration
AT dongsuyu maetransrnetanimprovedtransformerconvnetarchitecturewithmaskedautoencoderforcardiacmriregistration
AT yuyang maetransrnetanimprovedtransformerconvnetarchitecturewithmaskedautoencoderforcardiacmriregistration
AT liyan maetransrnetanimprovedtransformerconvnetarchitecturewithmaskedautoencoderforcardiacmriregistration
AT yangguangyuan maetransrnetanimprovedtransformerconvnetarchitecturewithmaskedautoencoderforcardiacmriregistration
AT qiuzhaowen maetransrnetanimprovedtransformerconvnetarchitecturewithmaskedautoencoderforcardiacmriregistration