Cargando…

A Manifold Learning Perspective on Representation Learning: Learning Decoder and Representations without an Encoder

Autoencoders are commonly used in representation learning. They consist of an encoder and a decoder, which provide a straightforward method to map n-dimensional data in input space to a lower m-dimensional representation space and back. The decoder itself defines an m-dimensional manifold in input s...

Descripción completa

Detalles Bibliográficos
Autores principales: Schuster, Viktoria, Krogh, Anders
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8625121/
https://www.ncbi.nlm.nih.gov/pubmed/34828101
http://dx.doi.org/10.3390/e23111403
_version_ 1784606341691604992
author Schuster, Viktoria
Krogh, Anders
author_facet Schuster, Viktoria
Krogh, Anders
author_sort Schuster, Viktoria
collection PubMed
description Autoencoders are commonly used in representation learning. They consist of an encoder and a decoder, which provide a straightforward method to map n-dimensional data in input space to a lower m-dimensional representation space and back. The decoder itself defines an m-dimensional manifold in input space. Inspired by manifold learning, we showed that the decoder can be trained on its own by learning the representations of the training samples along with the decoder weights using gradient descent. A sum-of-squares loss then corresponds to optimizing the manifold to have the smallest Euclidean distance to the training samples, and similarly for other loss functions. We derived expressions for the number of samples needed to specify the encoder and decoder and showed that the decoder generally requires much fewer training samples to be well-specified compared to the encoder. We discuss the training of autoencoders in this perspective and relate it to previous work in the field that uses noisy training examples and other types of regularization. On the natural image data sets MNIST and CIFAR10, we demonstrated that the decoder is much better suited to learn a low-dimensional representation, especially when trained on small data sets. Using simulated gene regulatory data, we further showed that the decoder alone leads to better generalization and meaningful representations. Our approach of training the decoder alone facilitates representation learning even on small data sets and can lead to improved training of autoencoders. We hope that the simple analyses presented will also contribute to an improved conceptual understanding of representation learning.
format Online
Article
Text
id pubmed-8625121
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-86251212021-11-27 A Manifold Learning Perspective on Representation Learning: Learning Decoder and Representations without an Encoder Schuster, Viktoria Krogh, Anders Entropy (Basel) Article Autoencoders are commonly used in representation learning. They consist of an encoder and a decoder, which provide a straightforward method to map n-dimensional data in input space to a lower m-dimensional representation space and back. The decoder itself defines an m-dimensional manifold in input space. Inspired by manifold learning, we showed that the decoder can be trained on its own by learning the representations of the training samples along with the decoder weights using gradient descent. A sum-of-squares loss then corresponds to optimizing the manifold to have the smallest Euclidean distance to the training samples, and similarly for other loss functions. We derived expressions for the number of samples needed to specify the encoder and decoder and showed that the decoder generally requires much fewer training samples to be well-specified compared to the encoder. We discuss the training of autoencoders in this perspective and relate it to previous work in the field that uses noisy training examples and other types of regularization. On the natural image data sets MNIST and CIFAR10, we demonstrated that the decoder is much better suited to learn a low-dimensional representation, especially when trained on small data sets. Using simulated gene regulatory data, we further showed that the decoder alone leads to better generalization and meaningful representations. Our approach of training the decoder alone facilitates representation learning even on small data sets and can lead to improved training of autoencoders. We hope that the simple analyses presented will also contribute to an improved conceptual understanding of representation learning. MDPI 2021-10-25 /pmc/articles/PMC8625121/ /pubmed/34828101 http://dx.doi.org/10.3390/e23111403 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Schuster, Viktoria
Krogh, Anders
A Manifold Learning Perspective on Representation Learning: Learning Decoder and Representations without an Encoder
title A Manifold Learning Perspective on Representation Learning: Learning Decoder and Representations without an Encoder
title_full A Manifold Learning Perspective on Representation Learning: Learning Decoder and Representations without an Encoder
title_fullStr A Manifold Learning Perspective on Representation Learning: Learning Decoder and Representations without an Encoder
title_full_unstemmed A Manifold Learning Perspective on Representation Learning: Learning Decoder and Representations without an Encoder
title_short A Manifold Learning Perspective on Representation Learning: Learning Decoder and Representations without an Encoder
title_sort manifold learning perspective on representation learning: learning decoder and representations without an encoder
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8625121/
https://www.ncbi.nlm.nih.gov/pubmed/34828101
http://dx.doi.org/10.3390/e23111403
work_keys_str_mv AT schusterviktoria amanifoldlearningperspectiveonrepresentationlearninglearningdecoderandrepresentationswithoutanencoder
AT kroghanders amanifoldlearningperspectiveonrepresentationlearninglearningdecoderandrepresentationswithoutanencoder
AT schusterviktoria manifoldlearningperspectiveonrepresentationlearninglearningdecoderandrepresentationswithoutanencoder
AT kroghanders manifoldlearningperspectiveonrepresentationlearninglearningdecoderandrepresentationswithoutanencoder