Corpus ID: 115855993

Ordering, Slicing And Splitting Monte Carlo Markov Chains

  title={Ordering, Slicing And Splitting Monte Carlo Markov Chains},
  author={Antonietta Mira},
  • A. Mira
  • Published 1998
  • Computer Science
Markov chain Monte Carlo is a method of approximating the integral of a function f with respect to a distribution . A Markov chain that has as its stationary distribution is simulated producing samplesX1; X2; : : : . The integral is approximated by taking the average of f(Xn) over the sample path. The standard way to construct such Markov chains is the Metropolis-Hastings algorithm. The class P of all Markov chains having as their unique stationary distribution is very large, so it is important… Expand
On the use of auxiliary variables in Markov chain Monte Carlo sampling
The slice sampler, a method of constructing a reversible Markov chain with a speciied invariant distribution, has a smaller second-largest eigenvalue than the corresponding independence Metropolis-Hastings algorithm. Expand
On the use of auxiliary variables inMarkov chain
We study the slice sampler, a method of constructing a reversible Markov chain with a speciied invariant distribution. Given an independence Metropolis-Hastings algorithm it is always possible toExpand
Slice Sampling 1
Markov chain sampling methods that adapt to characteristics of the distribution being sampled can be constructed using the principle that one can sample from a distribution by sampling uniformly fromExpand
Slice Sampling
Markov chain sampling methods that automatically adapt to characteristics of the distribution being sampled can be constructed by exploiting the principle that one can sample from a distribution byExpand
Limit theorems for sequential MCMC methods
Abstract Both sequential Monte Carlo (SMC) methods (a.k.a. ‘particle filters’) and sequential Markov chain Monte Carlo (sequential MCMC) methods constitute classes of algorithms which can be used toExpand
On extended state-space constructions for Monte Carlo methods
This thesis develops computationally efficient methodology in two areas. Firstly, we consider a particularly challenging class of discretely observed continuous-time point-process models. For these,Expand
Acknowledgments. This work is part of my doctoral research done under the direction of Jeerey S. Rosenthal. I thank Peter Rosenthal for helpful discussions about the operator theory issues. Abstract.Expand
Towards Automatic Reversible Jump Markov Chain Monte Carlo
The automatic sampler that is introduced in the penultimate chapter of the thesis builds upon the first steps taken by Green (2003) and uses adaptive techniques to perform self-tuning and calibration for many trans-dimensional statistical problems. Expand
Delayed Rejection in Reversible
In a Metropolis-Hastings algorithm, rejection of proposed moves is an intrinsic part of ensuring that the chain converges to the intended target distribution. However, persistent rejection, perhapsExpand
Delayed rejection Hamiltonian Monte Carlo for sampling multiscale distributions
A delayed rejection variant of Hamiltonian Monte Carlo that makes one or more subsequent proposals each using a step size geometrically smaller than the last if an initial HMC trajectory is rejected, providing increased robustness to step size misspecification. Expand


Markov Chain Monte Carlo Methods Based on `Slicing' the Density Function
Slice sampling is attractive in routine Markov chain Monte Carlo applications, and for use by software that automatically generates aMarkov chain sampler from a model specii-cation. Expand
Optimum Monte-Carlo sampling using Markov chains
SUMMARY The sampling method proposed by Metropolis et al. (1953) requires the simulation of a Markov chain with a specified 7i as its stationary distribution. Hastings (1970) outlined a generalExpand
Exact sampling with coupled Markov chains and applications to statistical mechanics
This work describes a simple variant of the Markov chain method that determines on its own when to stop, and that outputs samples in exact accordance with the desired distribution, and can sample from the Gibbs distributions associated with various statistical mechanics models. Expand
Practical Markov Chain Monte Carlo
The case is made for basing all inference on one long run of the Markov chain and estimating the Monte Carlo error by standard nonparametric methods well-known in the time-series and operations research literature. Expand
Suppressing Random Walks in Markov Chain Monte Carlo Using Ordered Overrelaxation
  • R. Neal
  • Mathematics, Physics
  • Learning in Graphical Models
  • 1998
An overrelaxed Markov chain Monte Carlo algorithm based on order statistics that can be applied whenever the full conditional distributions are such that their cumulative distribution functions and inverse cumulative distribution function can be efficiently computed. Expand
An interruptible algorithm for perfect sampling via Markov chains
A new algorithm is presented which again uses the same Markov chains to produce perfect samples from n, but is baaed on a different idea (namely, acceptance/rejection sampling); and eliminates user-impatience bias. Expand
Markov Chains for Exploring Posterior Distributions
Several Markov chain methods are available for sampling from a posterior distribution. Two important examples are the Gibbs sampler and the Metropolis algorithm. In addition, several strategies areExpand
We analyze the convergence to stationarity of a simple nonreversible Markov chain that serves as a model for several nonreversible Markov chain sampling methods that are used in practice. OurExpand
Metropolized independent sampling with comparisons to rejection sampling and importance sampling
  • J. Liu
  • Mathematics, Computer Science
  • Stat. Comput.
  • 1996
In this paper, a special Metropolis-Hastings algorithm, Metropolized independent sampling, proposed first in Hastings (1970), is studied in full detail and shown to be superior to rejection sampling in two respects: asymptotic efficiency and ease of computation. Expand
Auxiliary Variable Methods for Markov Chain Monte Carlo with Applications
Abstract Suppose that one wishes to sample from the density π(x) using Markov chain Monte Carlo (MCMC). An auxiliary variable u and its conditional distribution π(u|x) can be defined, giving theExpand