Cargando…

Gradient-based MCMC samplers for dynamic causal modelling

In this technical note, we derive two MCMC (Markov chain Monte Carlo) samplers for dynamic causal models (DCMs). Specifically, we use (a) Hamiltonian MCMC (HMC-E) where sampling is simulated using Hamilton’s equation of motion and (b) Langevin Monte Carlo algorithm (LMC-R and LMC-E) that simulates t...

Descripción completa

Detalles Bibliográficos
Autores principales: Sengupta, Biswa, Friston, Karl J., Penny, Will D.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Academic Press 2016
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4692453/
https://www.ncbi.nlm.nih.gov/pubmed/26213349
http://dx.doi.org/10.1016/j.neuroimage.2015.07.043
Descripción
Sumario:In this technical note, we derive two MCMC (Markov chain Monte Carlo) samplers for dynamic causal models (DCMs). Specifically, we use (a) Hamiltonian MCMC (HMC-E) where sampling is simulated using Hamilton’s equation of motion and (b) Langevin Monte Carlo algorithm (LMC-R and LMC-E) that simulates the Langevin diffusion of samples using gradients either on a Euclidean (E) or on a Riemannian (R) manifold. While LMC-R requires minimal tuning, the implementation of HMC-E is heavily dependent on its tuning parameters. These parameters are therefore optimised by learning a Gaussian process model of the time-normalised sample correlation matrix. This allows one to formulate an objective function that balances tuning parameter exploration and exploitation, furnishing an intervention-free inference scheme. Using neural mass models (NMMs)—a class of biophysically motivated DCMs—we find that HMC-E is statistically more efficient than LMC-R (with a Riemannian metric); yet both gradient-based samplers are far superior to the random walk Metropolis algorithm, which proves inadequate to steer away from dynamical instability.