Cargando…

Global Optimization Employing Gaussian Process-Based Bayesian Surrogates†

The simulation of complex physics models may lead to enormous computer running times. Since the simulations are expensive it is necessary to exploit the computational budget in the best possible manner. If for a few input parameter settings an output data set has been acquired, one could be interest...

Descripción completa

Detalles Bibliográficos
Autores principales: Preuss, Roland, von Toussaint, Udo
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2018
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7512716/
https://www.ncbi.nlm.nih.gov/pubmed/33265292
http://dx.doi.org/10.3390/e20030201
_version_ 1783586222066630656
author Preuss, Roland
von Toussaint, Udo
author_facet Preuss, Roland
von Toussaint, Udo
author_sort Preuss, Roland
collection PubMed
description The simulation of complex physics models may lead to enormous computer running times. Since the simulations are expensive it is necessary to exploit the computational budget in the best possible manner. If for a few input parameter settings an output data set has been acquired, one could be interested in taking these data as a basis for finding an extremum and possibly an input parameter set for further computer simulations to determine it—a task which belongs to the realm of global optimization. Within the Bayesian framework we utilize Gaussian processes for the creation of a surrogate model function adjusted self-consistently via hyperparameters to represent the data. Although the probability distribution of the hyperparameters may be widely spread over phase space, we make the assumption that only the use of their expectation values is sufficient. While this shortcut facilitates a quickly accessible surrogate, it is somewhat justified by the fact that we are not interested in a full representation of the model by the surrogate but to reveal its maximum. To accomplish this the surrogate is fed to a utility function whose extremum determines the new parameter set for the next data point to obtain. Moreover, we propose to alternate between two utility functions—expected improvement and maximum variance—in order to avoid the drawbacks of each. Subsequent data points are drawn from the model function until the procedure either remains in the points found or the surrogate model does not change with the iteration. The procedure is applied to mock data in one and two dimensions in order to demonstrate proof of principle of the proposed approach.
format Online
Article
Text
id pubmed-7512716
institution National Center for Biotechnology Information
language English
publishDate 2018
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-75127162020-11-09 Global Optimization Employing Gaussian Process-Based Bayesian Surrogates† Preuss, Roland von Toussaint, Udo Entropy (Basel) Article The simulation of complex physics models may lead to enormous computer running times. Since the simulations are expensive it is necessary to exploit the computational budget in the best possible manner. If for a few input parameter settings an output data set has been acquired, one could be interested in taking these data as a basis for finding an extremum and possibly an input parameter set for further computer simulations to determine it—a task which belongs to the realm of global optimization. Within the Bayesian framework we utilize Gaussian processes for the creation of a surrogate model function adjusted self-consistently via hyperparameters to represent the data. Although the probability distribution of the hyperparameters may be widely spread over phase space, we make the assumption that only the use of their expectation values is sufficient. While this shortcut facilitates a quickly accessible surrogate, it is somewhat justified by the fact that we are not interested in a full representation of the model by the surrogate but to reveal its maximum. To accomplish this the surrogate is fed to a utility function whose extremum determines the new parameter set for the next data point to obtain. Moreover, we propose to alternate between two utility functions—expected improvement and maximum variance—in order to avoid the drawbacks of each. Subsequent data points are drawn from the model function until the procedure either remains in the points found or the surrogate model does not change with the iteration. The procedure is applied to mock data in one and two dimensions in order to demonstrate proof of principle of the proposed approach. MDPI 2018-03-16 /pmc/articles/PMC7512716/ /pubmed/33265292 http://dx.doi.org/10.3390/e20030201 Text en © 2018 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Preuss, Roland
von Toussaint, Udo
Global Optimization Employing Gaussian Process-Based Bayesian Surrogates†
title Global Optimization Employing Gaussian Process-Based Bayesian Surrogates†
title_full Global Optimization Employing Gaussian Process-Based Bayesian Surrogates†
title_fullStr Global Optimization Employing Gaussian Process-Based Bayesian Surrogates†
title_full_unstemmed Global Optimization Employing Gaussian Process-Based Bayesian Surrogates†
title_short Global Optimization Employing Gaussian Process-Based Bayesian Surrogates†
title_sort global optimization employing gaussian process-based bayesian surrogates†
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7512716/
https://www.ncbi.nlm.nih.gov/pubmed/33265292
http://dx.doi.org/10.3390/e20030201
work_keys_str_mv AT preussroland globaloptimizationemployinggaussianprocessbasedbayesiansurrogates
AT vontoussaintudo globaloptimizationemployinggaussianprocessbasedbayesiansurrogates