Cargando…

Reconstructing seen images from human brain activity via guided stochastic search

Visual reconstruction algorithms are an interpretive tool that map brain activity to pixels. Past reconstruction algorithms employed brute-force search through a massive library to select candidate images that, when passed through an encoding model, accurately predict brain activity. Here, we use co...

Descripción completa

Detalles Bibliográficos
Autores principales: Kneeland, Reese, Ojeda, Jordyn, St-Yves, Ghislain, Naselaris, Thomas
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Cornell University 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10187366/
https://www.ncbi.nlm.nih.gov/pubmed/37205268
_version_ 1785042725971689472
author Kneeland, Reese
Ojeda, Jordyn
St-Yves, Ghislain
Naselaris, Thomas
author_facet Kneeland, Reese
Ojeda, Jordyn
St-Yves, Ghislain
Naselaris, Thomas
author_sort Kneeland, Reese
collection PubMed
description Visual reconstruction algorithms are an interpretive tool that map brain activity to pixels. Past reconstruction algorithms employed brute-force search through a massive library to select candidate images that, when passed through an encoding model, accurately predict brain activity. Here, we use conditional generative diffusion models to extend and improve this search-based strategy. We decode a semantic descriptor from human brain activity (7T fMRI) in voxels across most of visual cortex, then use a diffusion model to sample a small library of images conditioned on this descriptor. We pass each sample through an encoding model, select the images that best predict brain activity, and then use these images to seed another library. We show that this process converges on high-quality reconstructions by refining low-level image details while preserving semantic content across iterations. Interestingly, the time-to-convergence differs systematically across visual cortex, suggesting a succinct new way to measure the diversity of representations across visual brain areas.
format Online
Article
Text
id pubmed-10187366
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Cornell University
record_format MEDLINE/PubMed
spelling pubmed-101873662023-05-17 Reconstructing seen images from human brain activity via guided stochastic search Kneeland, Reese Ojeda, Jordyn St-Yves, Ghislain Naselaris, Thomas ArXiv Article Visual reconstruction algorithms are an interpretive tool that map brain activity to pixels. Past reconstruction algorithms employed brute-force search through a massive library to select candidate images that, when passed through an encoding model, accurately predict brain activity. Here, we use conditional generative diffusion models to extend and improve this search-based strategy. We decode a semantic descriptor from human brain activity (7T fMRI) in voxels across most of visual cortex, then use a diffusion model to sample a small library of images conditioned on this descriptor. We pass each sample through an encoding model, select the images that best predict brain activity, and then use these images to seed another library. We show that this process converges on high-quality reconstructions by refining low-level image details while preserving semantic content across iterations. Interestingly, the time-to-convergence differs systematically across visual cortex, suggesting a succinct new way to measure the diversity of representations across visual brain areas. Cornell University 2023-05-02 /pmc/articles/PMC10187366/ /pubmed/37205268 Text en https://creativecommons.org/licenses/by-nc-nd/4.0/This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License (https://creativecommons.org/licenses/by-nc-nd/4.0/) , which allows reusers to copy and distribute the material in any medium or format in unadapted form only, for noncommercial purposes only, and only so long as attribution is given to the creator.
spellingShingle Article
Kneeland, Reese
Ojeda, Jordyn
St-Yves, Ghislain
Naselaris, Thomas
Reconstructing seen images from human brain activity via guided stochastic search
title Reconstructing seen images from human brain activity via guided stochastic search
title_full Reconstructing seen images from human brain activity via guided stochastic search
title_fullStr Reconstructing seen images from human brain activity via guided stochastic search
title_full_unstemmed Reconstructing seen images from human brain activity via guided stochastic search
title_short Reconstructing seen images from human brain activity via guided stochastic search
title_sort reconstructing seen images from human brain activity via guided stochastic search
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10187366/
https://www.ncbi.nlm.nih.gov/pubmed/37205268
work_keys_str_mv AT kneelandreese reconstructingseenimagesfromhumanbrainactivityviaguidedstochasticsearch
AT ojedajordyn reconstructingseenimagesfromhumanbrainactivityviaguidedstochasticsearch
AT styvesghislain reconstructingseenimagesfromhumanbrainactivityviaguidedstochasticsearch
AT naselaristhomas reconstructingseenimagesfromhumanbrainactivityviaguidedstochasticsearch