Cargando…

Adversarially Training MCMC with Non-Volume-Preserving Flows

Recently, flow models parameterized by neural networks have been used to design efficient Markov chain Monte Carlo (MCMC) transition kernels. However, inefficient utilization of gradient information of the target distribution or the use of volume-preserving flows limits their performance in sampling...

Descripción completa

Detalles Bibliográficos
Autores principales: Liu, Shaofan, Sun, Shiliang
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8947447/
https://www.ncbi.nlm.nih.gov/pubmed/35327925
http://dx.doi.org/10.3390/e24030415
_version_ 1784674441393864704
author Liu, Shaofan
Sun, Shiliang
author_facet Liu, Shaofan
Sun, Shiliang
author_sort Liu, Shaofan
collection PubMed
description Recently, flow models parameterized by neural networks have been used to design efficient Markov chain Monte Carlo (MCMC) transition kernels. However, inefficient utilization of gradient information of the target distribution or the use of volume-preserving flows limits their performance in sampling from multi-modal target distributions. In this paper, we treat the training procedure of the parameterized transition kernels in a different manner and exploit a novel scheme to train MCMC transition kernels. We divide the training process of transition kernels into the exploration stage and training stage, which can make full use of the gradient information of the target distribution and the expressive power of deep neural networks. The transition kernels are constructed with non-volume-preserving flows and trained in an adversarial form. The proposed method achieves significant improvement in effective sample size and mixes quickly to the target distribution. Empirical results validate that the proposed method is able to achieve low autocorrelation of samples and fast convergence rates, and outperforms other state-of-the-art parameterized transition kernels in varieties of challenging analytically described distributions and real world datasets.
format Online
Article
Text
id pubmed-8947447
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-89474472022-03-25 Adversarially Training MCMC with Non-Volume-Preserving Flows Liu, Shaofan Sun, Shiliang Entropy (Basel) Article Recently, flow models parameterized by neural networks have been used to design efficient Markov chain Monte Carlo (MCMC) transition kernels. However, inefficient utilization of gradient information of the target distribution or the use of volume-preserving flows limits their performance in sampling from multi-modal target distributions. In this paper, we treat the training procedure of the parameterized transition kernels in a different manner and exploit a novel scheme to train MCMC transition kernels. We divide the training process of transition kernels into the exploration stage and training stage, which can make full use of the gradient information of the target distribution and the expressive power of deep neural networks. The transition kernels are constructed with non-volume-preserving flows and trained in an adversarial form. The proposed method achieves significant improvement in effective sample size and mixes quickly to the target distribution. Empirical results validate that the proposed method is able to achieve low autocorrelation of samples and fast convergence rates, and outperforms other state-of-the-art parameterized transition kernels in varieties of challenging analytically described distributions and real world datasets. MDPI 2022-03-16 /pmc/articles/PMC8947447/ /pubmed/35327925 http://dx.doi.org/10.3390/e24030415 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Liu, Shaofan
Sun, Shiliang
Adversarially Training MCMC with Non-Volume-Preserving Flows
title Adversarially Training MCMC with Non-Volume-Preserving Flows
title_full Adversarially Training MCMC with Non-Volume-Preserving Flows
title_fullStr Adversarially Training MCMC with Non-Volume-Preserving Flows
title_full_unstemmed Adversarially Training MCMC with Non-Volume-Preserving Flows
title_short Adversarially Training MCMC with Non-Volume-Preserving Flows
title_sort adversarially training mcmc with non-volume-preserving flows
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8947447/
https://www.ncbi.nlm.nih.gov/pubmed/35327925
http://dx.doi.org/10.3390/e24030415
work_keys_str_mv AT liushaofan adversariallytrainingmcmcwithnonvolumepreservingflows
AT sunshiliang adversariallytrainingmcmcwithnonvolumepreservingflows