Cargando…
Conditional Deep Gaussian Processes: Empirical Bayes Hyperdata Learning
It is desirable to combine the expressive power of deep learning with Gaussian Process (GP) in one expressive Bayesian learning model. Deep kernel learning showed success as a deep network used for feature extraction. Then, a GP was used as the function model. Recently, it was suggested that, albeit...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8618322/ https://www.ncbi.nlm.nih.gov/pubmed/34828085 http://dx.doi.org/10.3390/e23111387 |
_version_ | 1784604719630516224 |
---|---|
author | Lu, Chi-Ken Shafto, Patrick |
author_facet | Lu, Chi-Ken Shafto, Patrick |
author_sort | Lu, Chi-Ken |
collection | PubMed |
description | It is desirable to combine the expressive power of deep learning with Gaussian Process (GP) in one expressive Bayesian learning model. Deep kernel learning showed success as a deep network used for feature extraction. Then, a GP was used as the function model. Recently, it was suggested that, albeit training with marginal likelihood, the deterministic nature of a feature extractor might lead to overfitting, and replacement with a Bayesian network seemed to cure it. Here, we propose the conditional deep Gaussian process (DGP) in which the intermediate GPs in hierarchical composition are supported by the hyperdata and the exposed GP remains zero mean. Motivated by the inducing points in sparse GP, the hyperdata also play the role of function supports, but are hyperparameters rather than random variables. It follows our previous moment matching approach to approximate the marginal prior for conditional DGP with a GP carrying an effective kernel. Thus, as in empirical Bayes, the hyperdata are learned by optimizing the approximate marginal likelihood which implicitly depends on the hyperdata via the kernel. We show the equivalence with the deep kernel learning in the limit of dense hyperdata in latent space. However, the conditional DGP and the corresponding approximate inference enjoy the benefit of being more Bayesian than deep kernel learning. Preliminary extrapolation results demonstrate expressive power from the depth of hierarchy by exploiting the exact covariance and hyperdata learning, in comparison with GP kernel composition, DGP variational inference and deep kernel learning. We also address the non-Gaussian aspect of our model as well as way of upgrading to a full Bayes inference. |
format | Online Article Text |
id | pubmed-8618322 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-86183222021-11-27 Conditional Deep Gaussian Processes: Empirical Bayes Hyperdata Learning Lu, Chi-Ken Shafto, Patrick Entropy (Basel) Article It is desirable to combine the expressive power of deep learning with Gaussian Process (GP) in one expressive Bayesian learning model. Deep kernel learning showed success as a deep network used for feature extraction. Then, a GP was used as the function model. Recently, it was suggested that, albeit training with marginal likelihood, the deterministic nature of a feature extractor might lead to overfitting, and replacement with a Bayesian network seemed to cure it. Here, we propose the conditional deep Gaussian process (DGP) in which the intermediate GPs in hierarchical composition are supported by the hyperdata and the exposed GP remains zero mean. Motivated by the inducing points in sparse GP, the hyperdata also play the role of function supports, but are hyperparameters rather than random variables. It follows our previous moment matching approach to approximate the marginal prior for conditional DGP with a GP carrying an effective kernel. Thus, as in empirical Bayes, the hyperdata are learned by optimizing the approximate marginal likelihood which implicitly depends on the hyperdata via the kernel. We show the equivalence with the deep kernel learning in the limit of dense hyperdata in latent space. However, the conditional DGP and the corresponding approximate inference enjoy the benefit of being more Bayesian than deep kernel learning. Preliminary extrapolation results demonstrate expressive power from the depth of hierarchy by exploiting the exact covariance and hyperdata learning, in comparison with GP kernel composition, DGP variational inference and deep kernel learning. We also address the non-Gaussian aspect of our model as well as way of upgrading to a full Bayes inference. MDPI 2021-10-23 /pmc/articles/PMC8618322/ /pubmed/34828085 http://dx.doi.org/10.3390/e23111387 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Lu, Chi-Ken Shafto, Patrick Conditional Deep Gaussian Processes: Empirical Bayes Hyperdata Learning |
title | Conditional Deep Gaussian Processes: Empirical Bayes Hyperdata Learning |
title_full | Conditional Deep Gaussian Processes: Empirical Bayes Hyperdata Learning |
title_fullStr | Conditional Deep Gaussian Processes: Empirical Bayes Hyperdata Learning |
title_full_unstemmed | Conditional Deep Gaussian Processes: Empirical Bayes Hyperdata Learning |
title_short | Conditional Deep Gaussian Processes: Empirical Bayes Hyperdata Learning |
title_sort | conditional deep gaussian processes: empirical bayes hyperdata learning |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8618322/ https://www.ncbi.nlm.nih.gov/pubmed/34828085 http://dx.doi.org/10.3390/e23111387 |
work_keys_str_mv | AT luchiken conditionaldeepgaussianprocessesempiricalbayeshyperdatalearning AT shaftopatrick conditionaldeepgaussianprocessesempiricalbayeshyperdatalearning |