Cargando…

Zero-shot visual reasoning through probabilistic analogical mapping

Human reasoning is grounded in an ability to identify highly abstract commonalities governing superficially dissimilar visual inputs. Recent efforts to develop algorithms with this capacity have largely focused on approaches that require extensive direct training on visual reasoning tasks, and yield...

Descripción completa

Detalles Bibliográficos
Autores principales: Webb, Taylor, Fu, Shuhao, Bihl, Trevor, Holyoak, Keith J., Lu, Hongjing
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Nature Publishing Group UK 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10449798/
https://www.ncbi.nlm.nih.gov/pubmed/37620313
http://dx.doi.org/10.1038/s41467-023-40804-x
_version_ 1785095039034064896
author Webb, Taylor
Fu, Shuhao
Bihl, Trevor
Holyoak, Keith J.
Lu, Hongjing
author_facet Webb, Taylor
Fu, Shuhao
Bihl, Trevor
Holyoak, Keith J.
Lu, Hongjing
author_sort Webb, Taylor
collection PubMed
description Human reasoning is grounded in an ability to identify highly abstract commonalities governing superficially dissimilar visual inputs. Recent efforts to develop algorithms with this capacity have largely focused on approaches that require extensive direct training on visual reasoning tasks, and yield limited generalization to problems with novel content. In contrast, a long tradition of research in cognitive science has focused on elucidating the computational principles underlying human analogical reasoning; however, this work has generally relied on manually constructed representations. Here we present visiPAM (visual Probabilistic Analogical Mapping), a model of visual reasoning that synthesizes these two approaches. VisiPAM employs learned representations derived directly from naturalistic visual inputs, coupled with a similarity-based mapping operation derived from cognitive theories of human reasoning. We show that without any direct training, visiPAM outperforms a state-of-the-art deep learning model on an analogical mapping task. In addition, visiPAM closely matches the pattern of human performance on a novel task involving mapping of 3D objects across disparate categories.
format Online
Article
Text
id pubmed-10449798
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Nature Publishing Group UK
record_format MEDLINE/PubMed
spelling pubmed-104497982023-08-26 Zero-shot visual reasoning through probabilistic analogical mapping Webb, Taylor Fu, Shuhao Bihl, Trevor Holyoak, Keith J. Lu, Hongjing Nat Commun Article Human reasoning is grounded in an ability to identify highly abstract commonalities governing superficially dissimilar visual inputs. Recent efforts to develop algorithms with this capacity have largely focused on approaches that require extensive direct training on visual reasoning tasks, and yield limited generalization to problems with novel content. In contrast, a long tradition of research in cognitive science has focused on elucidating the computational principles underlying human analogical reasoning; however, this work has generally relied on manually constructed representations. Here we present visiPAM (visual Probabilistic Analogical Mapping), a model of visual reasoning that synthesizes these two approaches. VisiPAM employs learned representations derived directly from naturalistic visual inputs, coupled with a similarity-based mapping operation derived from cognitive theories of human reasoning. We show that without any direct training, visiPAM outperforms a state-of-the-art deep learning model on an analogical mapping task. In addition, visiPAM closely matches the pattern of human performance on a novel task involving mapping of 3D objects across disparate categories. Nature Publishing Group UK 2023-08-24 /pmc/articles/PMC10449798/ /pubmed/37620313 http://dx.doi.org/10.1038/s41467-023-40804-x Text en © The Author(s) 2023 https://creativecommons.org/licenses/by/4.0/Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) .
spellingShingle Article
Webb, Taylor
Fu, Shuhao
Bihl, Trevor
Holyoak, Keith J.
Lu, Hongjing
Zero-shot visual reasoning through probabilistic analogical mapping
title Zero-shot visual reasoning through probabilistic analogical mapping
title_full Zero-shot visual reasoning through probabilistic analogical mapping
title_fullStr Zero-shot visual reasoning through probabilistic analogical mapping
title_full_unstemmed Zero-shot visual reasoning through probabilistic analogical mapping
title_short Zero-shot visual reasoning through probabilistic analogical mapping
title_sort zero-shot visual reasoning through probabilistic analogical mapping
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10449798/
https://www.ncbi.nlm.nih.gov/pubmed/37620313
http://dx.doi.org/10.1038/s41467-023-40804-x
work_keys_str_mv AT webbtaylor zeroshotvisualreasoningthroughprobabilisticanalogicalmapping
AT fushuhao zeroshotvisualreasoningthroughprobabilisticanalogicalmapping
AT bihltrevor zeroshotvisualreasoningthroughprobabilisticanalogicalmapping
AT holyoakkeithj zeroshotvisualreasoningthroughprobabilisticanalogicalmapping
AT luhongjing zeroshotvisualreasoningthroughprobabilisticanalogicalmapping