Abstract

Slice sampling has emerged as a powerful Markov Chain Monte Carlo algorithm that adapts to the characteristics of the target distribution with minimal hand-tuning. However, Slice Sampling’s performance is highly sensitive to the user-specified initial length scale hyperparameter and the method generally struggles with poorly scaled or strongly correlated distributions. This paper introduces Ensemble Slice Sampling (ESS), a new class of algorithms that bypasses such difficulties by adaptively tuning the initial length scale and utilising an ensemble of parallel walkers in order to efficiently handle strong correlations between parameters. These affine-invariant algorithms are trivial to construct, require no hand-tuning, and can easily be implemented in parallel computing environments. Empirical tests show that Ensemble Slice Sampling can improve efficiency by more than an order of magnitude compared to conventional MCMC methods on a broad range of highly correlated target distributions. In cases of strongly multimodal target distributions, Ensemble Slice Sampling can sample efficiently even in high dimensions. We argue that the parallel, black-box and gradient-free nature of the method renders it ideal for use in scientific fields such as physics, astrophysics and cosmology which are dominated by a wide variety of computationally expensive and non-differentiable models.

Highlights

  • Bayesian inference and data analysis has become an integral part of modern science

  • This need led to the development of adaptive Markov Chain Monte Carlo (MCMC) methods like the Adaptive Metropolis algorithm (Haario et al 2001) which tunes its proposal scale based on the sample covariance matrix

  • We found that the computational overhead introduced by the variational fitting of the Dirchlet process Gaussian Mixture (DPGM) is negligible compared to the computational cost of the evaluation of the model and posterior distribution in common problems in physics, astrophysics and cosmology

Read more

Summary

Introduction

Bayesian inference and data analysis has become an integral part of modern science This is partly due to the ability of Markov Chain Monte Carlo (MCMC) algorithms to generate samples from intractable probability distributions. The emerging and routine use of such mathematical tools in science calls for the development of black-box MCMC algorithms that require no hand-tuning at all This need led to the development of adaptive MCMC methods like the Adaptive Metropolis algorithm (Haario et al 2001) which tunes its proposal scale based on the sample covariance matrix. There is no reason to believe that a single Metropolis proposal scale is optimal for the whole distribution (i.e. the appropriate scale could vary from one part of the distribution to another) Another approach to deal with those issues would be to develop methods that by construction require no or minimal hand-tuning

61 Page 2 of 18
Standard Slice Sampling
Adaptively tuning the length scale
Ensemble Slice Sampling
61 Page 4 of 18
Ensemble of walkers
Affine transformations and invariance
Differential move
Gaussian move
Global move
Parallelising the ensemble
61 Page 8 of 18
Empirical evaluation
Performance tests
Correlated funnel
61 Page 10 of 18
Comparison to other ensemble methods
Ring distribution
Gaussian shells distribution
Hierarchical Gaussian process regression
Bayesian object detection
Gaussian Mixture
Convergence of the length scale
Parallel scaling
Discussion
61 Page 16 of 18
A Estimating the effective sample size
61 Page 18 of 18
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call