Cargando…
A Visual Encoding Model Based on Contrastive Self-Supervised Learning for Human Brain Activity along the Ventral Visual Stream
Visual encoding models are important computational models for understanding how information is processed along the visual stream. Many improved visual encoding models have been developed from the perspective of the model architecture and the learning objective, but these are limited to the supervise...
Autores principales: | , , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8391143/ https://www.ncbi.nlm.nih.gov/pubmed/34439623 http://dx.doi.org/10.3390/brainsci11081004 |
_version_ | 1783743201834696704 |
---|---|
author | Li, Jingwei Zhang, Chi Wang, Linyuan Ding, Penghui Hu, Lulu Yan, Bin Tong, Li |
author_facet | Li, Jingwei Zhang, Chi Wang, Linyuan Ding, Penghui Hu, Lulu Yan, Bin Tong, Li |
author_sort | Li, Jingwei |
collection | PubMed |
description | Visual encoding models are important computational models for understanding how information is processed along the visual stream. Many improved visual encoding models have been developed from the perspective of the model architecture and the learning objective, but these are limited to the supervised learning method. From the view of unsupervised learning mechanisms, this paper utilized a pre-trained neural network to construct a visual encoding model based on contrastive self-supervised learning for the ventral visual stream measured by functional magnetic resonance imaging (fMRI). We first extracted features using the ResNet50 model pre-trained in contrastive self-supervised learning (ResNet50-CSL model), trained a linear regression model for each voxel, and finally calculated the prediction accuracy of different voxels. Compared with the ResNet50 model pre-trained in a supervised classification task, the ResNet50-CSL model achieved an equal or even relatively better encoding performance in multiple visual cortical areas. Moreover, the ResNet50-CSL model performs hierarchical representation of input visual stimuli, which is similar to the human visual cortex in its hierarchical information processing. Our experimental results suggest that the encoding model based on contrastive self-supervised learning is a strong computational model to compete with supervised models, and contrastive self-supervised learning proves an effective learning method to extract human brain-like representations. |
format | Online Article Text |
id | pubmed-8391143 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-83911432021-08-28 A Visual Encoding Model Based on Contrastive Self-Supervised Learning for Human Brain Activity along the Ventral Visual Stream Li, Jingwei Zhang, Chi Wang, Linyuan Ding, Penghui Hu, Lulu Yan, Bin Tong, Li Brain Sci Article Visual encoding models are important computational models for understanding how information is processed along the visual stream. Many improved visual encoding models have been developed from the perspective of the model architecture and the learning objective, but these are limited to the supervised learning method. From the view of unsupervised learning mechanisms, this paper utilized a pre-trained neural network to construct a visual encoding model based on contrastive self-supervised learning for the ventral visual stream measured by functional magnetic resonance imaging (fMRI). We first extracted features using the ResNet50 model pre-trained in contrastive self-supervised learning (ResNet50-CSL model), trained a linear regression model for each voxel, and finally calculated the prediction accuracy of different voxels. Compared with the ResNet50 model pre-trained in a supervised classification task, the ResNet50-CSL model achieved an equal or even relatively better encoding performance in multiple visual cortical areas. Moreover, the ResNet50-CSL model performs hierarchical representation of input visual stimuli, which is similar to the human visual cortex in its hierarchical information processing. Our experimental results suggest that the encoding model based on contrastive self-supervised learning is a strong computational model to compete with supervised models, and contrastive self-supervised learning proves an effective learning method to extract human brain-like representations. MDPI 2021-07-29 /pmc/articles/PMC8391143/ /pubmed/34439623 http://dx.doi.org/10.3390/brainsci11081004 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Li, Jingwei Zhang, Chi Wang, Linyuan Ding, Penghui Hu, Lulu Yan, Bin Tong, Li A Visual Encoding Model Based on Contrastive Self-Supervised Learning for Human Brain Activity along the Ventral Visual Stream |
title | A Visual Encoding Model Based on Contrastive Self-Supervised Learning for Human Brain Activity along the Ventral Visual Stream |
title_full | A Visual Encoding Model Based on Contrastive Self-Supervised Learning for Human Brain Activity along the Ventral Visual Stream |
title_fullStr | A Visual Encoding Model Based on Contrastive Self-Supervised Learning for Human Brain Activity along the Ventral Visual Stream |
title_full_unstemmed | A Visual Encoding Model Based on Contrastive Self-Supervised Learning for Human Brain Activity along the Ventral Visual Stream |
title_short | A Visual Encoding Model Based on Contrastive Self-Supervised Learning for Human Brain Activity along the Ventral Visual Stream |
title_sort | visual encoding model based on contrastive self-supervised learning for human brain activity along the ventral visual stream |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8391143/ https://www.ncbi.nlm.nih.gov/pubmed/34439623 http://dx.doi.org/10.3390/brainsci11081004 |
work_keys_str_mv | AT lijingwei avisualencodingmodelbasedoncontrastiveselfsupervisedlearningforhumanbrainactivityalongtheventralvisualstream AT zhangchi avisualencodingmodelbasedoncontrastiveselfsupervisedlearningforhumanbrainactivityalongtheventralvisualstream AT wanglinyuan avisualencodingmodelbasedoncontrastiveselfsupervisedlearningforhumanbrainactivityalongtheventralvisualstream AT dingpenghui avisualencodingmodelbasedoncontrastiveselfsupervisedlearningforhumanbrainactivityalongtheventralvisualstream AT hululu avisualencodingmodelbasedoncontrastiveselfsupervisedlearningforhumanbrainactivityalongtheventralvisualstream AT yanbin avisualencodingmodelbasedoncontrastiveselfsupervisedlearningforhumanbrainactivityalongtheventralvisualstream AT tongli avisualencodingmodelbasedoncontrastiveselfsupervisedlearningforhumanbrainactivityalongtheventralvisualstream AT lijingwei visualencodingmodelbasedoncontrastiveselfsupervisedlearningforhumanbrainactivityalongtheventralvisualstream AT zhangchi visualencodingmodelbasedoncontrastiveselfsupervisedlearningforhumanbrainactivityalongtheventralvisualstream AT wanglinyuan visualencodingmodelbasedoncontrastiveselfsupervisedlearningforhumanbrainactivityalongtheventralvisualstream AT dingpenghui visualencodingmodelbasedoncontrastiveselfsupervisedlearningforhumanbrainactivityalongtheventralvisualstream AT hululu visualencodingmodelbasedoncontrastiveselfsupervisedlearningforhumanbrainactivityalongtheventralvisualstream AT yanbin visualencodingmodelbasedoncontrastiveselfsupervisedlearningforhumanbrainactivityalongtheventralvisualstream AT tongli visualencodingmodelbasedoncontrastiveselfsupervisedlearningforhumanbrainactivityalongtheventralvisualstream |