Cargando…

Modeling language and cognition with deep unsupervised learning: a tutorial overview

Deep unsupervised learning in stochastic recurrent neural networks with many layers of hidden units is a recent breakthrough in neural computation research. These networks build a hierarchy of progressively more complex distributed representations of the sensory data by fitting a hierarchical genera...

Descripción completa

Detalles Bibliográficos
Autores principales: Zorzi, Marco, Testolin, Alberto, Stoianov, Ivilin P.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2013
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3747356/
https://www.ncbi.nlm.nih.gov/pubmed/23970869
http://dx.doi.org/10.3389/fpsyg.2013.00515
_version_ 1782280916315406336
author Zorzi, Marco
Testolin, Alberto
Stoianov, Ivilin P.
author_facet Zorzi, Marco
Testolin, Alberto
Stoianov, Ivilin P.
author_sort Zorzi, Marco
collection PubMed
description Deep unsupervised learning in stochastic recurrent neural networks with many layers of hidden units is a recent breakthrough in neural computation research. These networks build a hierarchy of progressively more complex distributed representations of the sensory data by fitting a hierarchical generative model. In this article we discuss the theoretical foundations of this approach and we review key issues related to training, testing and analysis of deep networks for modeling language and cognitive processing. The classic letter and word perception problem of McClelland and Rumelhart (1981) is used as a tutorial example to illustrate how structured and abstract representations may emerge from deep generative learning. We argue that the focus on deep architectures and generative (rather than discriminative) learning represents a crucial step forward for the connectionist modeling enterprise, because it offers a more plausible model of cortical learning as well as a way to bridge the gap between emergentist connectionist models and structured Bayesian models of cognition.
format Online
Article
Text
id pubmed-3747356
institution National Center for Biotechnology Information
language English
publishDate 2013
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-37473562013-08-22 Modeling language and cognition with deep unsupervised learning: a tutorial overview Zorzi, Marco Testolin, Alberto Stoianov, Ivilin P. Front Psychol Psychology Deep unsupervised learning in stochastic recurrent neural networks with many layers of hidden units is a recent breakthrough in neural computation research. These networks build a hierarchy of progressively more complex distributed representations of the sensory data by fitting a hierarchical generative model. In this article we discuss the theoretical foundations of this approach and we review key issues related to training, testing and analysis of deep networks for modeling language and cognitive processing. The classic letter and word perception problem of McClelland and Rumelhart (1981) is used as a tutorial example to illustrate how structured and abstract representations may emerge from deep generative learning. We argue that the focus on deep architectures and generative (rather than discriminative) learning represents a crucial step forward for the connectionist modeling enterprise, because it offers a more plausible model of cortical learning as well as a way to bridge the gap between emergentist connectionist models and structured Bayesian models of cognition. Frontiers Media S.A. 2013-08-20 /pmc/articles/PMC3747356/ /pubmed/23970869 http://dx.doi.org/10.3389/fpsyg.2013.00515 Text en Copyright © 2013 Zorzi, Testolin and Stoianov. http://creativecommons.org/licenses/by/3.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Psychology
Zorzi, Marco
Testolin, Alberto
Stoianov, Ivilin P.
Modeling language and cognition with deep unsupervised learning: a tutorial overview
title Modeling language and cognition with deep unsupervised learning: a tutorial overview
title_full Modeling language and cognition with deep unsupervised learning: a tutorial overview
title_fullStr Modeling language and cognition with deep unsupervised learning: a tutorial overview
title_full_unstemmed Modeling language and cognition with deep unsupervised learning: a tutorial overview
title_short Modeling language and cognition with deep unsupervised learning: a tutorial overview
title_sort modeling language and cognition with deep unsupervised learning: a tutorial overview
topic Psychology
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3747356/
https://www.ncbi.nlm.nih.gov/pubmed/23970869
http://dx.doi.org/10.3389/fpsyg.2013.00515
work_keys_str_mv AT zorzimarco modelinglanguageandcognitionwithdeepunsupervisedlearningatutorialoverview
AT testolinalberto modelinglanguageandcognitionwithdeepunsupervisedlearningatutorialoverview
AT stoianovivilinp modelinglanguageandcognitionwithdeepunsupervisedlearningatutorialoverview