Cargando…

Conditional Deep Gaussian Processes: Multi-Fidelity Kernel Learning

Deep Gaussian Processes (DGPs) were proposed as an expressive Bayesian model capable of a mathematically grounded estimation of uncertainty. The expressivity of DPGs results from not only the compositional character but the distribution propagation within the hierarchy. Recently, it was pointed out...

Descripción completa

Detalles Bibliográficos
Autores principales: Lu, Chi-Ken, Shafto, Patrick
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8625033/
https://www.ncbi.nlm.nih.gov/pubmed/34828243
http://dx.doi.org/10.3390/e23111545
_version_ 1784606319104229376
author Lu, Chi-Ken
Shafto, Patrick
author_facet Lu, Chi-Ken
Shafto, Patrick
author_sort Lu, Chi-Ken
collection PubMed
description Deep Gaussian Processes (DGPs) were proposed as an expressive Bayesian model capable of a mathematically grounded estimation of uncertainty. The expressivity of DPGs results from not only the compositional character but the distribution propagation within the hierarchy. Recently, it was pointed out that the hierarchical structure of DGP well suited modeling the multi-fidelity regression, in which one is provided sparse observations with high precision and plenty of low fidelity observations. We propose the conditional DGP model in which the latent GPs are directly supported by the fixed lower fidelity data. Then the moment matching method is applied to approximate the marginal prior of conditional DGP with a GP. The obtained effective kernels are implicit functions of the lower-fidelity data, manifesting the expressivity contributed by distribution propagation within the hierarchy. The hyperparameters are learned via optimizing the approximate marginal likelihood. Experiments with synthetic and high dimensional data show comparable performance against other multi-fidelity regression methods, variational inference, and multi-output GP. We conclude that, with the low fidelity data and the hierarchical DGP structure, the effective kernel encodes the inductive bias for true function allowing the compositional freedom.
format Online
Article
Text
id pubmed-8625033
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-86250332021-11-27 Conditional Deep Gaussian Processes: Multi-Fidelity Kernel Learning Lu, Chi-Ken Shafto, Patrick Entropy (Basel) Article Deep Gaussian Processes (DGPs) were proposed as an expressive Bayesian model capable of a mathematically grounded estimation of uncertainty. The expressivity of DPGs results from not only the compositional character but the distribution propagation within the hierarchy. Recently, it was pointed out that the hierarchical structure of DGP well suited modeling the multi-fidelity regression, in which one is provided sparse observations with high precision and plenty of low fidelity observations. We propose the conditional DGP model in which the latent GPs are directly supported by the fixed lower fidelity data. Then the moment matching method is applied to approximate the marginal prior of conditional DGP with a GP. The obtained effective kernels are implicit functions of the lower-fidelity data, manifesting the expressivity contributed by distribution propagation within the hierarchy. The hyperparameters are learned via optimizing the approximate marginal likelihood. Experiments with synthetic and high dimensional data show comparable performance against other multi-fidelity regression methods, variational inference, and multi-output GP. We conclude that, with the low fidelity data and the hierarchical DGP structure, the effective kernel encodes the inductive bias for true function allowing the compositional freedom. MDPI 2021-11-20 /pmc/articles/PMC8625033/ /pubmed/34828243 http://dx.doi.org/10.3390/e23111545 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Lu, Chi-Ken
Shafto, Patrick
Conditional Deep Gaussian Processes: Multi-Fidelity Kernel Learning
title Conditional Deep Gaussian Processes: Multi-Fidelity Kernel Learning
title_full Conditional Deep Gaussian Processes: Multi-Fidelity Kernel Learning
title_fullStr Conditional Deep Gaussian Processes: Multi-Fidelity Kernel Learning
title_full_unstemmed Conditional Deep Gaussian Processes: Multi-Fidelity Kernel Learning
title_short Conditional Deep Gaussian Processes: Multi-Fidelity Kernel Learning
title_sort conditional deep gaussian processes: multi-fidelity kernel learning
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8625033/
https://www.ncbi.nlm.nih.gov/pubmed/34828243
http://dx.doi.org/10.3390/e23111545
work_keys_str_mv AT luchiken conditionaldeepgaussianprocessesmultifidelitykernellearning
AT shaftopatrick conditionaldeepgaussianprocessesmultifidelitykernellearning