Cargando…
Zero-shot visual reasoning through probabilistic analogical mapping
Human reasoning is grounded in an ability to identify highly abstract commonalities governing superficially dissimilar visual inputs. Recent efforts to develop algorithms with this capacity have largely focused on approaches that require extensive direct training on visual reasoning tasks, and yield...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Nature Publishing Group UK
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10449798/ https://www.ncbi.nlm.nih.gov/pubmed/37620313 http://dx.doi.org/10.1038/s41467-023-40804-x |
_version_ | 1785095039034064896 |
---|---|
author | Webb, Taylor Fu, Shuhao Bihl, Trevor Holyoak, Keith J. Lu, Hongjing |
author_facet | Webb, Taylor Fu, Shuhao Bihl, Trevor Holyoak, Keith J. Lu, Hongjing |
author_sort | Webb, Taylor |
collection | PubMed |
description | Human reasoning is grounded in an ability to identify highly abstract commonalities governing superficially dissimilar visual inputs. Recent efforts to develop algorithms with this capacity have largely focused on approaches that require extensive direct training on visual reasoning tasks, and yield limited generalization to problems with novel content. In contrast, a long tradition of research in cognitive science has focused on elucidating the computational principles underlying human analogical reasoning; however, this work has generally relied on manually constructed representations. Here we present visiPAM (visual Probabilistic Analogical Mapping), a model of visual reasoning that synthesizes these two approaches. VisiPAM employs learned representations derived directly from naturalistic visual inputs, coupled with a similarity-based mapping operation derived from cognitive theories of human reasoning. We show that without any direct training, visiPAM outperforms a state-of-the-art deep learning model on an analogical mapping task. In addition, visiPAM closely matches the pattern of human performance on a novel task involving mapping of 3D objects across disparate categories. |
format | Online Article Text |
id | pubmed-10449798 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Nature Publishing Group UK |
record_format | MEDLINE/PubMed |
spelling | pubmed-104497982023-08-26 Zero-shot visual reasoning through probabilistic analogical mapping Webb, Taylor Fu, Shuhao Bihl, Trevor Holyoak, Keith J. Lu, Hongjing Nat Commun Article Human reasoning is grounded in an ability to identify highly abstract commonalities governing superficially dissimilar visual inputs. Recent efforts to develop algorithms with this capacity have largely focused on approaches that require extensive direct training on visual reasoning tasks, and yield limited generalization to problems with novel content. In contrast, a long tradition of research in cognitive science has focused on elucidating the computational principles underlying human analogical reasoning; however, this work has generally relied on manually constructed representations. Here we present visiPAM (visual Probabilistic Analogical Mapping), a model of visual reasoning that synthesizes these two approaches. VisiPAM employs learned representations derived directly from naturalistic visual inputs, coupled with a similarity-based mapping operation derived from cognitive theories of human reasoning. We show that without any direct training, visiPAM outperforms a state-of-the-art deep learning model on an analogical mapping task. In addition, visiPAM closely matches the pattern of human performance on a novel task involving mapping of 3D objects across disparate categories. Nature Publishing Group UK 2023-08-24 /pmc/articles/PMC10449798/ /pubmed/37620313 http://dx.doi.org/10.1038/s41467-023-40804-x Text en © The Author(s) 2023 https://creativecommons.org/licenses/by/4.0/Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) . |
spellingShingle | Article Webb, Taylor Fu, Shuhao Bihl, Trevor Holyoak, Keith J. Lu, Hongjing Zero-shot visual reasoning through probabilistic analogical mapping |
title | Zero-shot visual reasoning through probabilistic analogical mapping |
title_full | Zero-shot visual reasoning through probabilistic analogical mapping |
title_fullStr | Zero-shot visual reasoning through probabilistic analogical mapping |
title_full_unstemmed | Zero-shot visual reasoning through probabilistic analogical mapping |
title_short | Zero-shot visual reasoning through probabilistic analogical mapping |
title_sort | zero-shot visual reasoning through probabilistic analogical mapping |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10449798/ https://www.ncbi.nlm.nih.gov/pubmed/37620313 http://dx.doi.org/10.1038/s41467-023-40804-x |
work_keys_str_mv | AT webbtaylor zeroshotvisualreasoningthroughprobabilisticanalogicalmapping AT fushuhao zeroshotvisualreasoningthroughprobabilisticanalogicalmapping AT bihltrevor zeroshotvisualreasoningthroughprobabilisticanalogicalmapping AT holyoakkeithj zeroshotvisualreasoningthroughprobabilisticanalogicalmapping AT luhongjing zeroshotvisualreasoningthroughprobabilisticanalogicalmapping |