scispace - formally typeset
Open AccessProceedings Article

Estimating Convergence of Markov chains with L-Lag Couplings

Niloy Biswas, +2 more
- Vol. 32, pp 7389-7399
TLDR
In this paper, the authors introduce L-lag couplings to generate computable, nonasymptotic upper bound estimates for the total variation or the Wasserstein distance of general Markov chains.
Abstract
Markov chain Monte Carlo (MCMC) methods generate samples that are asymptotically distributed from a target distribution of interest as the number of iterations goes to infinity. Various theoretical results provide upper bounds on the distance between the target and marginal distribution after a fixed number of iterations. These upper bounds are on a case by case basis and typically involve intractable quantities, which limits their use for practitioners. We introduce L-lag couplings to generate computable, non-asymptotic upper bound estimates for the total variation or the Wasserstein distance of general Markov chains. We apply L-lag couplings to the tasks of (i) determining MCMC burn-in, (ii) comparing different MCMC algorithms with the same target, and (iii) comparing exact and approximate MCMC. Lastly, we (iv) assess the bias of sequential Monte Carlo and self-normalized importance samplers.

read more

Content maybe subject to copyright    Report

Citations
More filters
Journal ArticleDOI

Unbiased Markov chain Monte Carlo methods with couplings

TL;DR: The theoretical validity of the estimators proposed and their efficiency relative to the underlying MCMC algorithms are established and the performance and limitations of the method are illustrated.
Posted Content

Optimal Thinning of MCMC Output

TL;DR: A novel method is proposed, based on greedy minimisation of a kernel Stein discrepancy, that is suitable for problems where heavy compression is required and its effectiveness is demonstrated in the challenging context of parameter inference for ordinary differential equations.
Posted Content

An invitation to sequential Monte Carlo samplers

TL;DR: This article presents this class of methods and a number of recent advances, with the goal of helping statisticians assess the applicability and usefulness of these methods for their purposes.
Journal ArticleDOI

A Unified Framework for De-Duplication and Population Size Estimation (with Discussion)

TL;DR: This paper tackles the de-duplication process via a latent entity model, where the observed data are perturbed versions of a set of key variables drawn from a finite population of N different entities, to consider the population size N as an unknown model parameter.
Journal ArticleDOI

A Gibbs Sampler for a Class of Random Convex Polytopes

TL;DR: A Gibbs sampler for the Dempster–Shafer approach to statistical inference for categorical distributions is presented and relies on an equivalence between the iterative constraints of the vertex configuration and the nonnegativity of cycles in a fully connected directed graph.
References
More filters
Journal Article

R: A language and environment for statistical computing.

R Core Team
- 01 Jan 2014 - 
TL;DR: Copyright (©) 1999–2012 R Foundation for Statistical Computing; permission is granted to make and distribute verbatim copies of this manual provided the copyright notice and permission notice are preserved on all copies.
Book

ggplot2: Elegant Graphics for Data Analysis

TL;DR: This book describes ggplot2, a new data visualization package for R that uses the insights from Leland Wilkisons Grammar of Graphics to create a powerful and flexible system for creating data graphics.
Journal ArticleDOI

Inference from Iterative Simulation Using Multiple Sequences

TL;DR: The focus is on applied inference for Bayesian posterior distributions in real problems, which often tend toward normal- ity after transformations and marginalization, and the results are derived as normal-theory approximations to exact Bayesian inference, conditional on the observed simulations.
Journal ArticleDOI

General methods for monitoring convergence of iterative simulations

TL;DR: This work generalizes the method proposed by Gelman and Rubin (1992a) for monitoring the convergence of iterative simulations by comparing between and within variances of multiple chains, in order to obtain a family of tests for convergence.

Stan: A Probabilistic Programming Language.

TL;DR: Stan is a probabilistic programming language for specifying statistical models that provides full Bayesian inference for continuous-variable models through Markov chain Monte Carlo methods such as the No-U-Turn sampler and an adaptive form of Hamiltonian Monte Carlo sampling.
Related Papers (5)