Cargando…
Can deep convolutional neural networks support relational reasoning in the same-different task?
Same-different visual reasoning is a basic skill central to abstract combinatorial thought. This fact has lead neural networks researchers to test same-different classification on deep convolutional neural networks (DCNNs), which has resulted in a controversy regarding whether this skill is within t...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
The Association for Research in Vision and Ophthalmology
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9482325/ https://www.ncbi.nlm.nih.gov/pubmed/36094524 http://dx.doi.org/10.1167/jov.22.10.11 |
_version_ | 1784791432486191104 |
---|---|
author | Puebla, Guillermo Bowers, Jeffrey S. |
author_facet | Puebla, Guillermo Bowers, Jeffrey S. |
author_sort | Puebla, Guillermo |
collection | PubMed |
description | Same-different visual reasoning is a basic skill central to abstract combinatorial thought. This fact has lead neural networks researchers to test same-different classification on deep convolutional neural networks (DCNNs), which has resulted in a controversy regarding whether this skill is within the capacity of these models. However, most tests of same-different classification rely on testing on images that come from the same pixel-level distribution as the training images, yielding the results inconclusive. In this study, we tested relational same-different reasoning in DCNNs. In a series of simulations we show that models based on the ResNet architecture are capable of visual same-different classification, but only when the test images are similar to the training images at the pixel level. In contrast, when there is a shift in the testing distribution that does not change the relation between the objects in the image, the performance of DCNNs decreases substantially. This finding is true even when the DCNNs’ training regime is expanded to include images taken from a wide range of different pixel-level distributions or when the model is trained on the testing distribution but on a different task in a multitask learning context. Furthermore, we show that the relation network, a deep learning architecture specifically designed to tackle visual relational reasoning problems, suffers the same kind of limitations. Overall, the results of this study suggest that learning same-different relations is beyond the scope of current DCNNs. |
format | Online Article Text |
id | pubmed-9482325 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | The Association for Research in Vision and Ophthalmology |
record_format | MEDLINE/PubMed |
spelling | pubmed-94823252022-09-18 Can deep convolutional neural networks support relational reasoning in the same-different task? Puebla, Guillermo Bowers, Jeffrey S. J Vis Article Same-different visual reasoning is a basic skill central to abstract combinatorial thought. This fact has lead neural networks researchers to test same-different classification on deep convolutional neural networks (DCNNs), which has resulted in a controversy regarding whether this skill is within the capacity of these models. However, most tests of same-different classification rely on testing on images that come from the same pixel-level distribution as the training images, yielding the results inconclusive. In this study, we tested relational same-different reasoning in DCNNs. In a series of simulations we show that models based on the ResNet architecture are capable of visual same-different classification, but only when the test images are similar to the training images at the pixel level. In contrast, when there is a shift in the testing distribution that does not change the relation between the objects in the image, the performance of DCNNs decreases substantially. This finding is true even when the DCNNs’ training regime is expanded to include images taken from a wide range of different pixel-level distributions or when the model is trained on the testing distribution but on a different task in a multitask learning context. Furthermore, we show that the relation network, a deep learning architecture specifically designed to tackle visual relational reasoning problems, suffers the same kind of limitations. Overall, the results of this study suggest that learning same-different relations is beyond the scope of current DCNNs. The Association for Research in Vision and Ophthalmology 2022-09-12 /pmc/articles/PMC9482325/ /pubmed/36094524 http://dx.doi.org/10.1167/jov.22.10.11 Text en Copyright 2022 The Authors https://creativecommons.org/licenses/by-nc-nd/4.0/This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License. |
spellingShingle | Article Puebla, Guillermo Bowers, Jeffrey S. Can deep convolutional neural networks support relational reasoning in the same-different task? |
title | Can deep convolutional neural networks support relational reasoning in the same-different task? |
title_full | Can deep convolutional neural networks support relational reasoning in the same-different task? |
title_fullStr | Can deep convolutional neural networks support relational reasoning in the same-different task? |
title_full_unstemmed | Can deep convolutional neural networks support relational reasoning in the same-different task? |
title_short | Can deep convolutional neural networks support relational reasoning in the same-different task? |
title_sort | can deep convolutional neural networks support relational reasoning in the same-different task? |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9482325/ https://www.ncbi.nlm.nih.gov/pubmed/36094524 http://dx.doi.org/10.1167/jov.22.10.11 |
work_keys_str_mv | AT pueblaguillermo candeepconvolutionalneuralnetworkssupportrelationalreasoninginthesamedifferenttask AT bowersjeffreys candeepconvolutionalneuralnetworkssupportrelationalreasoninginthesamedifferenttask |