Cargando…

A multimodal fusion method for Alzheimer’s disease based on DCT convolutional sparse representation

INTRODUCTION: The medical information contained in magnetic resonance imaging (MRI) and positron emission tomography (PET) has driven the development of intelligent diagnosis of Alzheimer’s disease (AD) and multimodal medical imaging. To solve the problems of severe energy loss, low contrast of fuse...

Descripción completa

Detalles Bibliográficos
Autores principales: Zhang, Guo, Nie, Xixi, Liu, Bangtao, Yuan, Hong, Li, Jin, Sun, Weiwei, Huang, Shixin
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9853298/
https://www.ncbi.nlm.nih.gov/pubmed/36685238
http://dx.doi.org/10.3389/fnins.2022.1100812
Descripción
Sumario:INTRODUCTION: The medical information contained in magnetic resonance imaging (MRI) and positron emission tomography (PET) has driven the development of intelligent diagnosis of Alzheimer’s disease (AD) and multimodal medical imaging. To solve the problems of severe energy loss, low contrast of fused images and spatial inconsistency in the traditional multimodal medical image fusion methods based on sparse representation. A multimodal fusion algorithm for Alzheimer’ s disease based on the discrete cosine transform (DCT) convolutional sparse representation is proposed. METHODS: The algorithm first performs a multi-scale DCT decomposition of the source medical images and uses the sub-images of different scales as training images, respectively. Different sparse coefficients are obtained by optimally solving the sub-dictionaries at different scales using alternating directional multiplication method (ADMM). Secondly, the coefficients of high-frequency and low-frequency subimages are inverse DCTed using an improved L1 parametric rule combined with improved spatial frequency novel sum-modified SF (NMSF) to obtain the final fused images. RESULTS AND DISCUSSION: Through extensive experimental results, we show that our proposed method has good performance in contrast enhancement, texture and contour information retention.