Cargando…

Data-Dependent Conditional Priors for Unsupervised Learning of Multimodal Data †

One of the major shortcomings of variational autoencoders is the inability to produce generations from the individual modalities of data originating from mixture distributions. This is primarily due to the use of a simple isotropic Gaussian as the prior for the latent code in the ancestral sampling...

Descripción completa

Detalles Bibliográficos
Autores principales: Lavda, Frantzeska, Gregorová, Magda, Kalousis, Alexandros
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7517502/
https://www.ncbi.nlm.nih.gov/pubmed/33286658
http://dx.doi.org/10.3390/e22080888
_version_ 1783587240858877952
author Lavda, Frantzeska
Gregorová, Magda
Kalousis, Alexandros
author_facet Lavda, Frantzeska
Gregorová, Magda
Kalousis, Alexandros
author_sort Lavda, Frantzeska
collection PubMed
description One of the major shortcomings of variational autoencoders is the inability to produce generations from the individual modalities of data originating from mixture distributions. This is primarily due to the use of a simple isotropic Gaussian as the prior for the latent code in the ancestral sampling procedure for data generations. In this paper, we propose a novel formulation of variational autoencoders, conditional prior VAE (CP-VAE), with a two-level generative process for the observed data where continuous [Formula: see text] and a discrete [Formula: see text] variables are introduced in addition to the observed variables [Formula: see text]. By learning data-dependent conditional priors, the new variational objective naturally encourages a better match between the posterior and prior conditionals, and the learning of the latent categories encoding the major source of variation of the original data in an unsupervised manner. Through sampling continuous latent code from the data-dependent conditional priors, we are able to generate new samples from the individual mixture components corresponding, to the multimodal structure over the original data. Moreover, we unify and analyse our objective under different independence assumptions for the joint distribution of the continuous and discrete latent variables. We provide an empirical evaluation on one synthetic dataset and three image datasets, FashionMNIST, MNIST, and Omniglot, illustrating the generative performance of our new model comparing to multiple baselines.
format Online
Article
Text
id pubmed-7517502
institution National Center for Biotechnology Information
language English
publishDate 2020
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-75175022020-11-09 Data-Dependent Conditional Priors for Unsupervised Learning of Multimodal Data † Lavda, Frantzeska Gregorová, Magda Kalousis, Alexandros Entropy (Basel) Article One of the major shortcomings of variational autoencoders is the inability to produce generations from the individual modalities of data originating from mixture distributions. This is primarily due to the use of a simple isotropic Gaussian as the prior for the latent code in the ancestral sampling procedure for data generations. In this paper, we propose a novel formulation of variational autoencoders, conditional prior VAE (CP-VAE), with a two-level generative process for the observed data where continuous [Formula: see text] and a discrete [Formula: see text] variables are introduced in addition to the observed variables [Formula: see text]. By learning data-dependent conditional priors, the new variational objective naturally encourages a better match between the posterior and prior conditionals, and the learning of the latent categories encoding the major source of variation of the original data in an unsupervised manner. Through sampling continuous latent code from the data-dependent conditional priors, we are able to generate new samples from the individual mixture components corresponding, to the multimodal structure over the original data. Moreover, we unify and analyse our objective under different independence assumptions for the joint distribution of the continuous and discrete latent variables. We provide an empirical evaluation on one synthetic dataset and three image datasets, FashionMNIST, MNIST, and Omniglot, illustrating the generative performance of our new model comparing to multiple baselines. MDPI 2020-08-13 /pmc/articles/PMC7517502/ /pubmed/33286658 http://dx.doi.org/10.3390/e22080888 Text en © 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Lavda, Frantzeska
Gregorová, Magda
Kalousis, Alexandros
Data-Dependent Conditional Priors for Unsupervised Learning of Multimodal Data †
title Data-Dependent Conditional Priors for Unsupervised Learning of Multimodal Data †
title_full Data-Dependent Conditional Priors for Unsupervised Learning of Multimodal Data †
title_fullStr Data-Dependent Conditional Priors for Unsupervised Learning of Multimodal Data †
title_full_unstemmed Data-Dependent Conditional Priors for Unsupervised Learning of Multimodal Data †
title_short Data-Dependent Conditional Priors for Unsupervised Learning of Multimodal Data †
title_sort data-dependent conditional priors for unsupervised learning of multimodal data †
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7517502/
https://www.ncbi.nlm.nih.gov/pubmed/33286658
http://dx.doi.org/10.3390/e22080888
work_keys_str_mv AT lavdafrantzeska datadependentconditionalpriorsforunsupervisedlearningofmultimodaldata
AT gregorovamagda datadependentconditionalpriorsforunsupervisedlearningofmultimodaldata
AT kalousisalexandros datadependentconditionalpriorsforunsupervisedlearningofmultimodaldata