Cargando…
Deep generative learning of location-invariant visual word recognition
It is widely believed that orthographic processing implies an approximate, flexible coding of letter position, as shown by relative-position and transposition priming effects in visual word recognition. These findings have inspired alternative proposals about the representation of letter position, r...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2013
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3776941/ https://www.ncbi.nlm.nih.gov/pubmed/24065939 http://dx.doi.org/10.3389/fpsyg.2013.00635 |
_version_ | 1782284910773403648 |
---|---|
author | Di Bono, Maria Grazia Zorzi, Marco |
author_facet | Di Bono, Maria Grazia Zorzi, Marco |
author_sort | Di Bono, Maria Grazia |
collection | PubMed |
description | It is widely believed that orthographic processing implies an approximate, flexible coding of letter position, as shown by relative-position and transposition priming effects in visual word recognition. These findings have inspired alternative proposals about the representation of letter position, ranging from noisy coding across the ordinal positions to relative position coding based on open bigrams. This debate can be cast within the broader problem of learning location-invariant representations of written words, that is, a coding scheme abstracting the identity and position of letters (and combinations of letters) from their eye-centered (i.e., retinal) locations. We asked whether location-invariance would emerge from deep unsupervised learning on letter strings and what type of intermediate coding would emerge in the resulting hierarchical generative model. We trained a deep network with three hidden layers on an artificial dataset of letter strings presented at five possible retinal locations. Though word-level information (i.e., word identity) was never provided to the network during training, linear decoding from the activity of the deepest hidden layer yielded near-perfect accuracy in location-invariant word recognition. Conversely, decoding from lower layers yielded a large number of transposition errors. Analyses of emergent internal representations showed that word selectivity and location invariance increased as a function of layer depth. Word-tuning and location-invariance were found at the level of single neurons, but there was no evidence for bigram coding. Finally, the distributed internal representation of words at the deepest layer showed higher similarity to the representation elicited by the two exterior letters than by other combinations of two contiguous letters, in agreement with the hypothesis that word edges have special status. These results reveal that the efficient coding of written words—which was the model's learning objective—is largely based on letter-level information. |
format | Online Article Text |
id | pubmed-3776941 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2013 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-37769412013-09-24 Deep generative learning of location-invariant visual word recognition Di Bono, Maria Grazia Zorzi, Marco Front Psychol Psychology It is widely believed that orthographic processing implies an approximate, flexible coding of letter position, as shown by relative-position and transposition priming effects in visual word recognition. These findings have inspired alternative proposals about the representation of letter position, ranging from noisy coding across the ordinal positions to relative position coding based on open bigrams. This debate can be cast within the broader problem of learning location-invariant representations of written words, that is, a coding scheme abstracting the identity and position of letters (and combinations of letters) from their eye-centered (i.e., retinal) locations. We asked whether location-invariance would emerge from deep unsupervised learning on letter strings and what type of intermediate coding would emerge in the resulting hierarchical generative model. We trained a deep network with three hidden layers on an artificial dataset of letter strings presented at five possible retinal locations. Though word-level information (i.e., word identity) was never provided to the network during training, linear decoding from the activity of the deepest hidden layer yielded near-perfect accuracy in location-invariant word recognition. Conversely, decoding from lower layers yielded a large number of transposition errors. Analyses of emergent internal representations showed that word selectivity and location invariance increased as a function of layer depth. Word-tuning and location-invariance were found at the level of single neurons, but there was no evidence for bigram coding. Finally, the distributed internal representation of words at the deepest layer showed higher similarity to the representation elicited by the two exterior letters than by other combinations of two contiguous letters, in agreement with the hypothesis that word edges have special status. These results reveal that the efficient coding of written words—which was the model's learning objective—is largely based on letter-level information. Frontiers Media S.A. 2013-09-19 /pmc/articles/PMC3776941/ /pubmed/24065939 http://dx.doi.org/10.3389/fpsyg.2013.00635 Text en Copyright © 2013 Di Bono and Zorzi. http://creativecommons.org/licenses/by/3.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Psychology Di Bono, Maria Grazia Zorzi, Marco Deep generative learning of location-invariant visual word recognition |
title | Deep generative learning of location-invariant visual word recognition |
title_full | Deep generative learning of location-invariant visual word recognition |
title_fullStr | Deep generative learning of location-invariant visual word recognition |
title_full_unstemmed | Deep generative learning of location-invariant visual word recognition |
title_short | Deep generative learning of location-invariant visual word recognition |
title_sort | deep generative learning of location-invariant visual word recognition |
topic | Psychology |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3776941/ https://www.ncbi.nlm.nih.gov/pubmed/24065939 http://dx.doi.org/10.3389/fpsyg.2013.00635 |
work_keys_str_mv | AT dibonomariagrazia deepgenerativelearningoflocationinvariantvisualwordrecognition AT zorzimarco deepgenerativelearningoflocationinvariantvisualwordrecognition |