Convergence analysis of a collapsed Gibbs sampler for Bayesian vector autoregressions

KO Ekvall, GL Jones - 2021 - projecteuclid.org
2021projecteuclid.org
We study the convergence properties of a collapsed Gibbs sampler for Bayesian vector
autoregressions with predictors, or exogenous variables. The Markov chain generated by
our algorithm is shown to be geometrically ergodic regardless of whether the number of
observations in the underlying vector autoregression is small or large in comparison to the
order and dimension of it. In a convergence complexity analysis, we also give conditions for
when the geometric ergodicity is asymptotically stable as the number of observations tends …
Abstract
We study the convergence properties of a collapsed Gibbs sampler for Bayesian vector autoregressions with predictors, or exogenous variables. The Markov chain generated by our algorithm is shown to be geometrically ergodic regardless of whether the number of observations in the underlying vector autoregression is small or large in comparison to the order and dimension of it. In a convergence complexity analysis, we also give conditions for when the geometric ergodicity is asymptotically stable as the number of observations tends to infinity. Specifically, the geometric convergence rate is shown to be bounded away from unity asymptotically, either almost surely or with probability tending to one, depending on what is assumed about the data generating process. This result is one of the first of its kind for practically relevant Markov chain Monte Carlo algorithms. Our convergence results hold under close to arbitrary model misspecification.
Project Euclid