Cargando…

Can deep convolutional neural networks support relational reasoning in the same-different task?

Same-different visual reasoning is a basic skill central to abstract combinatorial thought. This fact has lead neural networks researchers to test same-different classification on deep convolutional neural networks (DCNNs), which has resulted in a controversy regarding whether this skill is within t...

Descripción completa

Detalles Bibliográficos
Autores principales: Puebla, Guillermo, Bowers, Jeffrey S.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: The Association for Research in Vision and Ophthalmology 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9482325/
https://www.ncbi.nlm.nih.gov/pubmed/36094524
http://dx.doi.org/10.1167/jov.22.10.11
_version_ 1784791432486191104
author Puebla, Guillermo
Bowers, Jeffrey S.
author_facet Puebla, Guillermo
Bowers, Jeffrey S.
author_sort Puebla, Guillermo
collection PubMed
description Same-different visual reasoning is a basic skill central to abstract combinatorial thought. This fact has lead neural networks researchers to test same-different classification on deep convolutional neural networks (DCNNs), which has resulted in a controversy regarding whether this skill is within the capacity of these models. However, most tests of same-different classification rely on testing on images that come from the same pixel-level distribution as the training images, yielding the results inconclusive. In this study, we tested relational same-different reasoning in DCNNs. In a series of simulations we show that models based on the ResNet architecture are capable of visual same-different classification, but only when the test images are similar to the training images at the pixel level. In contrast, when there is a shift in the testing distribution that does not change the relation between the objects in the image, the performance of DCNNs decreases substantially. This finding is true even when the DCNNs’ training regime is expanded to include images taken from a wide range of different pixel-level distributions or when the model is trained on the testing distribution but on a different task in a multitask learning context. Furthermore, we show that the relation network, a deep learning architecture specifically designed to tackle visual relational reasoning problems, suffers the same kind of limitations. Overall, the results of this study suggest that learning same-different relations is beyond the scope of current DCNNs.
format Online
Article
Text
id pubmed-9482325
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher The Association for Research in Vision and Ophthalmology
record_format MEDLINE/PubMed
spelling pubmed-94823252022-09-18 Can deep convolutional neural networks support relational reasoning in the same-different task? Puebla, Guillermo Bowers, Jeffrey S. J Vis Article Same-different visual reasoning is a basic skill central to abstract combinatorial thought. This fact has lead neural networks researchers to test same-different classification on deep convolutional neural networks (DCNNs), which has resulted in a controversy regarding whether this skill is within the capacity of these models. However, most tests of same-different classification rely on testing on images that come from the same pixel-level distribution as the training images, yielding the results inconclusive. In this study, we tested relational same-different reasoning in DCNNs. In a series of simulations we show that models based on the ResNet architecture are capable of visual same-different classification, but only when the test images are similar to the training images at the pixel level. In contrast, when there is a shift in the testing distribution that does not change the relation between the objects in the image, the performance of DCNNs decreases substantially. This finding is true even when the DCNNs’ training regime is expanded to include images taken from a wide range of different pixel-level distributions or when the model is trained on the testing distribution but on a different task in a multitask learning context. Furthermore, we show that the relation network, a deep learning architecture specifically designed to tackle visual relational reasoning problems, suffers the same kind of limitations. Overall, the results of this study suggest that learning same-different relations is beyond the scope of current DCNNs. The Association for Research in Vision and Ophthalmology 2022-09-12 /pmc/articles/PMC9482325/ /pubmed/36094524 http://dx.doi.org/10.1167/jov.22.10.11 Text en Copyright 2022 The Authors https://creativecommons.org/licenses/by-nc-nd/4.0/This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.
spellingShingle Article
Puebla, Guillermo
Bowers, Jeffrey S.
Can deep convolutional neural networks support relational reasoning in the same-different task?
title Can deep convolutional neural networks support relational reasoning in the same-different task?
title_full Can deep convolutional neural networks support relational reasoning in the same-different task?
title_fullStr Can deep convolutional neural networks support relational reasoning in the same-different task?
title_full_unstemmed Can deep convolutional neural networks support relational reasoning in the same-different task?
title_short Can deep convolutional neural networks support relational reasoning in the same-different task?
title_sort can deep convolutional neural networks support relational reasoning in the same-different task?
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9482325/
https://www.ncbi.nlm.nih.gov/pubmed/36094524
http://dx.doi.org/10.1167/jov.22.10.11
work_keys_str_mv AT pueblaguillermo candeepconvolutionalneuralnetworkssupportrelationalreasoninginthesamedifferenttask
AT bowersjeffreys candeepconvolutionalneuralnetworkssupportrelationalreasoninginthesamedifferenttask