Cargando…

Learning Numerosity Representations with Transformers: Number Generation Tasks and Out-of-Distribution Generalization

One of the most rapidly advancing areas of deep learning research aims at creating models that learn to disentangle the latent factors of variation from a data distribution. However, modeling joint probability mass functions is usually prohibitive, which motivates the use of conditional models assum...

Descripción completa

Detalles Bibliográficos
Autores principales: Boccato, Tommaso, Testolin, Alberto, Zorzi, Marco
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8303966/
https://www.ncbi.nlm.nih.gov/pubmed/34356398
http://dx.doi.org/10.3390/e23070857
_version_ 1783727218699010048
author Boccato, Tommaso
Testolin, Alberto
Zorzi, Marco
author_facet Boccato, Tommaso
Testolin, Alberto
Zorzi, Marco
author_sort Boccato, Tommaso
collection PubMed
description One of the most rapidly advancing areas of deep learning research aims at creating models that learn to disentangle the latent factors of variation from a data distribution. However, modeling joint probability mass functions is usually prohibitive, which motivates the use of conditional models assuming that some information is given as input. In the domain of numerical cognition, deep learning architectures have successfully demonstrated that approximate numerosity representations can emerge in multi-layer networks that build latent representations of a set of images with a varying number of items. However, existing models have focused on tasks requiring to conditionally estimate numerosity information from a given image. Here, we focus on a set of much more challenging tasks, which require to conditionally generate synthetic images containing a given number of items. We show that attention-based architectures operating at the pixel level can learn to produce well-formed images approximately containing a specific number of items, even when the target numerosity was not present in the training distribution.
format Online
Article
Text
id pubmed-8303966
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-83039662021-07-25 Learning Numerosity Representations with Transformers: Number Generation Tasks and Out-of-Distribution Generalization Boccato, Tommaso Testolin, Alberto Zorzi, Marco Entropy (Basel) Article One of the most rapidly advancing areas of deep learning research aims at creating models that learn to disentangle the latent factors of variation from a data distribution. However, modeling joint probability mass functions is usually prohibitive, which motivates the use of conditional models assuming that some information is given as input. In the domain of numerical cognition, deep learning architectures have successfully demonstrated that approximate numerosity representations can emerge in multi-layer networks that build latent representations of a set of images with a varying number of items. However, existing models have focused on tasks requiring to conditionally estimate numerosity information from a given image. Here, we focus on a set of much more challenging tasks, which require to conditionally generate synthetic images containing a given number of items. We show that attention-based architectures operating at the pixel level can learn to produce well-formed images approximately containing a specific number of items, even when the target numerosity was not present in the training distribution. MDPI 2021-07-03 /pmc/articles/PMC8303966/ /pubmed/34356398 http://dx.doi.org/10.3390/e23070857 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Boccato, Tommaso
Testolin, Alberto
Zorzi, Marco
Learning Numerosity Representations with Transformers: Number Generation Tasks and Out-of-Distribution Generalization
title Learning Numerosity Representations with Transformers: Number Generation Tasks and Out-of-Distribution Generalization
title_full Learning Numerosity Representations with Transformers: Number Generation Tasks and Out-of-Distribution Generalization
title_fullStr Learning Numerosity Representations with Transformers: Number Generation Tasks and Out-of-Distribution Generalization
title_full_unstemmed Learning Numerosity Representations with Transformers: Number Generation Tasks and Out-of-Distribution Generalization
title_short Learning Numerosity Representations with Transformers: Number Generation Tasks and Out-of-Distribution Generalization
title_sort learning numerosity representations with transformers: number generation tasks and out-of-distribution generalization
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8303966/
https://www.ncbi.nlm.nih.gov/pubmed/34356398
http://dx.doi.org/10.3390/e23070857
work_keys_str_mv AT boccatotommaso learningnumerosityrepresentationswithtransformersnumbergenerationtasksandoutofdistributiongeneralization
AT testolinalberto learningnumerosityrepresentationswithtransformersnumbergenerationtasksandoutofdistributiongeneralization
AT zorzimarco learningnumerosityrepresentationswithtransformersnumbergenerationtasksandoutofdistributiongeneralization