Cargando…
Data-Dependent Conditional Priors for Unsupervised Learning of Multimodal Data †
One of the major shortcomings of variational autoencoders is the inability to produce generations from the individual modalities of data originating from mixture distributions. This is primarily due to the use of a simple isotropic Gaussian as the prior for the latent code in the ancestral sampling...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2020
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7517502/ https://www.ncbi.nlm.nih.gov/pubmed/33286658 http://dx.doi.org/10.3390/e22080888 |
_version_ | 1783587240858877952 |
---|---|
author | Lavda, Frantzeska Gregorová, Magda Kalousis, Alexandros |
author_facet | Lavda, Frantzeska Gregorová, Magda Kalousis, Alexandros |
author_sort | Lavda, Frantzeska |
collection | PubMed |
description | One of the major shortcomings of variational autoencoders is the inability to produce generations from the individual modalities of data originating from mixture distributions. This is primarily due to the use of a simple isotropic Gaussian as the prior for the latent code in the ancestral sampling procedure for data generations. In this paper, we propose a novel formulation of variational autoencoders, conditional prior VAE (CP-VAE), with a two-level generative process for the observed data where continuous [Formula: see text] and a discrete [Formula: see text] variables are introduced in addition to the observed variables [Formula: see text]. By learning data-dependent conditional priors, the new variational objective naturally encourages a better match between the posterior and prior conditionals, and the learning of the latent categories encoding the major source of variation of the original data in an unsupervised manner. Through sampling continuous latent code from the data-dependent conditional priors, we are able to generate new samples from the individual mixture components corresponding, to the multimodal structure over the original data. Moreover, we unify and analyse our objective under different independence assumptions for the joint distribution of the continuous and discrete latent variables. We provide an empirical evaluation on one synthetic dataset and three image datasets, FashionMNIST, MNIST, and Omniglot, illustrating the generative performance of our new model comparing to multiple baselines. |
format | Online Article Text |
id | pubmed-7517502 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2020 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-75175022020-11-09 Data-Dependent Conditional Priors for Unsupervised Learning of Multimodal Data † Lavda, Frantzeska Gregorová, Magda Kalousis, Alexandros Entropy (Basel) Article One of the major shortcomings of variational autoencoders is the inability to produce generations from the individual modalities of data originating from mixture distributions. This is primarily due to the use of a simple isotropic Gaussian as the prior for the latent code in the ancestral sampling procedure for data generations. In this paper, we propose a novel formulation of variational autoencoders, conditional prior VAE (CP-VAE), with a two-level generative process for the observed data where continuous [Formula: see text] and a discrete [Formula: see text] variables are introduced in addition to the observed variables [Formula: see text]. By learning data-dependent conditional priors, the new variational objective naturally encourages a better match between the posterior and prior conditionals, and the learning of the latent categories encoding the major source of variation of the original data in an unsupervised manner. Through sampling continuous latent code from the data-dependent conditional priors, we are able to generate new samples from the individual mixture components corresponding, to the multimodal structure over the original data. Moreover, we unify and analyse our objective under different independence assumptions for the joint distribution of the continuous and discrete latent variables. We provide an empirical evaluation on one synthetic dataset and three image datasets, FashionMNIST, MNIST, and Omniglot, illustrating the generative performance of our new model comparing to multiple baselines. MDPI 2020-08-13 /pmc/articles/PMC7517502/ /pubmed/33286658 http://dx.doi.org/10.3390/e22080888 Text en © 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Lavda, Frantzeska Gregorová, Magda Kalousis, Alexandros Data-Dependent Conditional Priors for Unsupervised Learning of Multimodal Data † |
title | Data-Dependent Conditional Priors for Unsupervised Learning of Multimodal Data † |
title_full | Data-Dependent Conditional Priors for Unsupervised Learning of Multimodal Data † |
title_fullStr | Data-Dependent Conditional Priors for Unsupervised Learning of Multimodal Data † |
title_full_unstemmed | Data-Dependent Conditional Priors for Unsupervised Learning of Multimodal Data † |
title_short | Data-Dependent Conditional Priors for Unsupervised Learning of Multimodal Data † |
title_sort | data-dependent conditional priors for unsupervised learning of multimodal data † |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7517502/ https://www.ncbi.nlm.nih.gov/pubmed/33286658 http://dx.doi.org/10.3390/e22080888 |
work_keys_str_mv | AT lavdafrantzeska datadependentconditionalpriorsforunsupervisedlearningofmultimodaldata AT gregorovamagda datadependentconditionalpriorsforunsupervisedlearningofmultimodaldata AT kalousisalexandros datadependentconditionalpriorsforunsupervisedlearningofmultimodaldata |