Cargando…

Knot selection in sparse Gaussian processes with a variational objective function

Sparse, knot‐based Gaussian processes have enjoyed considerable success as scalable approximations of full Gaussian processes. Certain sparse models can be derived through specific variational approximations to the true posterior, and knots can be selected to minimize the Kullback‐Leibler divergence...

Descripción completa

Detalles Bibliográficos
Autores principales: Garton, Nathaniel, Niemi, Jarad, Carriquiry, Alicia
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Wiley Subscription Services, Inc., A Wiley Company 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7386924/
https://www.ncbi.nlm.nih.gov/pubmed/32742538
http://dx.doi.org/10.1002/sam.11459
Descripción
Sumario:Sparse, knot‐based Gaussian processes have enjoyed considerable success as scalable approximations of full Gaussian processes. Certain sparse models can be derived through specific variational approximations to the true posterior, and knots can be selected to minimize the Kullback‐Leibler divergence between the approximate and true posterior. While this has been a successful approach, simultaneous optimization of knots can be slow due to the number of parameters being optimized. Furthermore, there have been few proposed methods for selecting the number of knots, and no experimental results exist in the literature. We propose using a one‐at‐a‐time knot selection algorithm based on Bayesian optimization to select the number and locations of knots. We showcase the competitive performance of this method relative to optimization of knots simultaneously on three benchmark datasets, but at a fraction of the computational cost.