Cargando…

A transformer model for learning spatiotemporal contextual representation in fMRI data

Representation learning is a core component in data-driven modeling of various complex phenomena. Learning a contextually informative representation can especially benefit the analysis of fMRI data because of the complexities and dynamic dependencies present in such datasets. In this work, we propos...

Descripción completa

Detalles Bibliográficos
Autores principales: Asadi, Nima, Olson, Ingrid R., Obradovic, Zoran
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MIT Press 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10270708/
https://www.ncbi.nlm.nih.gov/pubmed/37334006
http://dx.doi.org/10.1162/netn_a_00281
_version_ 1785059371890245632
author Asadi, Nima
Olson, Ingrid R.
Obradovic, Zoran
author_facet Asadi, Nima
Olson, Ingrid R.
Obradovic, Zoran
author_sort Asadi, Nima
collection PubMed
description Representation learning is a core component in data-driven modeling of various complex phenomena. Learning a contextually informative representation can especially benefit the analysis of fMRI data because of the complexities and dynamic dependencies present in such datasets. In this work, we propose a framework based on transformer models to learn an embedding of the fMRI data by taking the spatiotemporal contextual information in the data into account. This approach takes the multivariate BOLD time series of the regions of the brain as well as their functional connectivity network simultaneously as the input to create a set of meaningful features that can in turn be used in various downstream tasks such as classification, feature extraction, and statistical analysis. The proposed spatiotemporal framework uses the attention mechanism as well as the graph convolution neural network to jointly inject the contextual information regarding the dynamics in time series data and their connectivity into the representation. We demonstrate the benefits of this framework by applying it to two resting-state fMRI datasets, and provide further discussion on various aspects and advantages of it over a number of other commonly adopted architectures.
format Online
Article
Text
id pubmed-10270708
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MIT Press
record_format MEDLINE/PubMed
spelling pubmed-102707082023-06-16 A transformer model for learning spatiotemporal contextual representation in fMRI data Asadi, Nima Olson, Ingrid R. Obradovic, Zoran Netw Neurosci Research Article Representation learning is a core component in data-driven modeling of various complex phenomena. Learning a contextually informative representation can especially benefit the analysis of fMRI data because of the complexities and dynamic dependencies present in such datasets. In this work, we propose a framework based on transformer models to learn an embedding of the fMRI data by taking the spatiotemporal contextual information in the data into account. This approach takes the multivariate BOLD time series of the regions of the brain as well as their functional connectivity network simultaneously as the input to create a set of meaningful features that can in turn be used in various downstream tasks such as classification, feature extraction, and statistical analysis. The proposed spatiotemporal framework uses the attention mechanism as well as the graph convolution neural network to jointly inject the contextual information regarding the dynamics in time series data and their connectivity into the representation. We demonstrate the benefits of this framework by applying it to two resting-state fMRI datasets, and provide further discussion on various aspects and advantages of it over a number of other commonly adopted architectures. MIT Press 2023-01-01 /pmc/articles/PMC10270708/ /pubmed/37334006 http://dx.doi.org/10.1162/netn_a_00281 Text en © 2022 Massachusetts Institute of Technology https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution 4.0 International License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. For a full description of the license, please visit https://creativecommons.org/licenses/by/4.0/.
spellingShingle Research Article
Asadi, Nima
Olson, Ingrid R.
Obradovic, Zoran
A transformer model for learning spatiotemporal contextual representation in fMRI data
title A transformer model for learning spatiotemporal contextual representation in fMRI data
title_full A transformer model for learning spatiotemporal contextual representation in fMRI data
title_fullStr A transformer model for learning spatiotemporal contextual representation in fMRI data
title_full_unstemmed A transformer model for learning spatiotemporal contextual representation in fMRI data
title_short A transformer model for learning spatiotemporal contextual representation in fMRI data
title_sort transformer model for learning spatiotemporal contextual representation in fmri data
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10270708/
https://www.ncbi.nlm.nih.gov/pubmed/37334006
http://dx.doi.org/10.1162/netn_a_00281
work_keys_str_mv AT asadinima atransformermodelforlearningspatiotemporalcontextualrepresentationinfmridata
AT olsoningridr atransformermodelforlearningspatiotemporalcontextualrepresentationinfmridata
AT obradoviczoran atransformermodelforlearningspatiotemporalcontextualrepresentationinfmridata
AT asadinima transformermodelforlearningspatiotemporalcontextualrepresentationinfmridata
AT olsoningridr transformermodelforlearningspatiotemporalcontextualrepresentationinfmridata
AT obradoviczoran transformermodelforlearningspatiotemporalcontextualrepresentationinfmridata