Cargando…

A Visual Encoding Model Based on Contrastive Self-Supervised Learning for Human Brain Activity along the Ventral Visual Stream

Visual encoding models are important computational models for understanding how information is processed along the visual stream. Many improved visual encoding models have been developed from the perspective of the model architecture and the learning objective, but these are limited to the supervise...

Descripción completa

Detalles Bibliográficos
Autores principales: Li, Jingwei, Zhang, Chi, Wang, Linyuan, Ding, Penghui, Hu, Lulu, Yan, Bin, Tong, Li
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8391143/
https://www.ncbi.nlm.nih.gov/pubmed/34439623
http://dx.doi.org/10.3390/brainsci11081004
_version_ 1783743201834696704
author Li, Jingwei
Zhang, Chi
Wang, Linyuan
Ding, Penghui
Hu, Lulu
Yan, Bin
Tong, Li
author_facet Li, Jingwei
Zhang, Chi
Wang, Linyuan
Ding, Penghui
Hu, Lulu
Yan, Bin
Tong, Li
author_sort Li, Jingwei
collection PubMed
description Visual encoding models are important computational models for understanding how information is processed along the visual stream. Many improved visual encoding models have been developed from the perspective of the model architecture and the learning objective, but these are limited to the supervised learning method. From the view of unsupervised learning mechanisms, this paper utilized a pre-trained neural network to construct a visual encoding model based on contrastive self-supervised learning for the ventral visual stream measured by functional magnetic resonance imaging (fMRI). We first extracted features using the ResNet50 model pre-trained in contrastive self-supervised learning (ResNet50-CSL model), trained a linear regression model for each voxel, and finally calculated the prediction accuracy of different voxels. Compared with the ResNet50 model pre-trained in a supervised classification task, the ResNet50-CSL model achieved an equal or even relatively better encoding performance in multiple visual cortical areas. Moreover, the ResNet50-CSL model performs hierarchical representation of input visual stimuli, which is similar to the human visual cortex in its hierarchical information processing. Our experimental results suggest that the encoding model based on contrastive self-supervised learning is a strong computational model to compete with supervised models, and contrastive self-supervised learning proves an effective learning method to extract human brain-like representations.
format Online
Article
Text
id pubmed-8391143
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-83911432021-08-28 A Visual Encoding Model Based on Contrastive Self-Supervised Learning for Human Brain Activity along the Ventral Visual Stream Li, Jingwei Zhang, Chi Wang, Linyuan Ding, Penghui Hu, Lulu Yan, Bin Tong, Li Brain Sci Article Visual encoding models are important computational models for understanding how information is processed along the visual stream. Many improved visual encoding models have been developed from the perspective of the model architecture and the learning objective, but these are limited to the supervised learning method. From the view of unsupervised learning mechanisms, this paper utilized a pre-trained neural network to construct a visual encoding model based on contrastive self-supervised learning for the ventral visual stream measured by functional magnetic resonance imaging (fMRI). We first extracted features using the ResNet50 model pre-trained in contrastive self-supervised learning (ResNet50-CSL model), trained a linear regression model for each voxel, and finally calculated the prediction accuracy of different voxels. Compared with the ResNet50 model pre-trained in a supervised classification task, the ResNet50-CSL model achieved an equal or even relatively better encoding performance in multiple visual cortical areas. Moreover, the ResNet50-CSL model performs hierarchical representation of input visual stimuli, which is similar to the human visual cortex in its hierarchical information processing. Our experimental results suggest that the encoding model based on contrastive self-supervised learning is a strong computational model to compete with supervised models, and contrastive self-supervised learning proves an effective learning method to extract human brain-like representations. MDPI 2021-07-29 /pmc/articles/PMC8391143/ /pubmed/34439623 http://dx.doi.org/10.3390/brainsci11081004 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Li, Jingwei
Zhang, Chi
Wang, Linyuan
Ding, Penghui
Hu, Lulu
Yan, Bin
Tong, Li
A Visual Encoding Model Based on Contrastive Self-Supervised Learning for Human Brain Activity along the Ventral Visual Stream
title A Visual Encoding Model Based on Contrastive Self-Supervised Learning for Human Brain Activity along the Ventral Visual Stream
title_full A Visual Encoding Model Based on Contrastive Self-Supervised Learning for Human Brain Activity along the Ventral Visual Stream
title_fullStr A Visual Encoding Model Based on Contrastive Self-Supervised Learning for Human Brain Activity along the Ventral Visual Stream
title_full_unstemmed A Visual Encoding Model Based on Contrastive Self-Supervised Learning for Human Brain Activity along the Ventral Visual Stream
title_short A Visual Encoding Model Based on Contrastive Self-Supervised Learning for Human Brain Activity along the Ventral Visual Stream
title_sort visual encoding model based on contrastive self-supervised learning for human brain activity along the ventral visual stream
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8391143/
https://www.ncbi.nlm.nih.gov/pubmed/34439623
http://dx.doi.org/10.3390/brainsci11081004
work_keys_str_mv AT lijingwei avisualencodingmodelbasedoncontrastiveselfsupervisedlearningforhumanbrainactivityalongtheventralvisualstream
AT zhangchi avisualencodingmodelbasedoncontrastiveselfsupervisedlearningforhumanbrainactivityalongtheventralvisualstream
AT wanglinyuan avisualencodingmodelbasedoncontrastiveselfsupervisedlearningforhumanbrainactivityalongtheventralvisualstream
AT dingpenghui avisualencodingmodelbasedoncontrastiveselfsupervisedlearningforhumanbrainactivityalongtheventralvisualstream
AT hululu avisualencodingmodelbasedoncontrastiveselfsupervisedlearningforhumanbrainactivityalongtheventralvisualstream
AT yanbin avisualencodingmodelbasedoncontrastiveselfsupervisedlearningforhumanbrainactivityalongtheventralvisualstream
AT tongli avisualencodingmodelbasedoncontrastiveselfsupervisedlearningforhumanbrainactivityalongtheventralvisualstream
AT lijingwei visualencodingmodelbasedoncontrastiveselfsupervisedlearningforhumanbrainactivityalongtheventralvisualstream
AT zhangchi visualencodingmodelbasedoncontrastiveselfsupervisedlearningforhumanbrainactivityalongtheventralvisualstream
AT wanglinyuan visualencodingmodelbasedoncontrastiveselfsupervisedlearningforhumanbrainactivityalongtheventralvisualstream
AT dingpenghui visualencodingmodelbasedoncontrastiveselfsupervisedlearningforhumanbrainactivityalongtheventralvisualstream
AT hululu visualencodingmodelbasedoncontrastiveselfsupervisedlearningforhumanbrainactivityalongtheventralvisualstream
AT yanbin visualencodingmodelbasedoncontrastiveselfsupervisedlearningforhumanbrainactivityalongtheventralvisualstream
AT tongli visualencodingmodelbasedoncontrastiveselfsupervisedlearningforhumanbrainactivityalongtheventralvisualstream