Corpus ID: 88514184

A fully Bayesian strategy for high-dimensional hierarchical modeling using massively parallel computing

  title={A fully Bayesian strategy for high-dimensional hierarchical modeling using massively parallel computing},
  author={Will Landau and Jarad Niemi},
  journal={arXiv: Computation},
Markov chain Monte Carlo (MCMC) is the predominant tool used in Bayesian parameter estimation for hierarchical models. When the model expands due to an increasing number of hierarchical levels, number of groups at a particular level, or number of observations in each group, a fully Bayesian analysis via MCMC can easily become computationally demanding, even intractable. We illustrate how the steps in an MCMC for hierarchical models are predominantly one of two types: conditionally independent… Expand

Figures and Tables from this paper

Applications of Bayesian hierarchical models in gene expression and product reliability
This work presents a Bayesian nonparametric hierarchical regression model for gene expression profiling data, and presents a method for estimation of lifetime for populations exhibiting heterogeneity due to infant mortality. Expand
Fully Bayesian Analysis of RNA-seq Counts for the Detection of Gene Expression Heterosis
A general hierarchical model for count data and a fully Bayesian analysis in which an efficient parallelized Markov chain Monte Carlo algorithm ameliorates the computational burden is developed and it is shown that hyperparameter posteriors are extremely narrow and an empirical Bayes (eBayes) approach provides virtually equivalent posterior probabilities, credible intervals, and gene rankings relative to the fullyBayesian solution. Expand
Bayesian analysis of high-dimensional count data
Diploid organisms have two copies of each gene (alleles) that can be separately transcribed. The RNA abundance of any particular allele is known as allele-specific expression (ASE). When two allelesExpand


Parallel MCMC with generalized elliptical slice sampling
A parallelizable Markov chain Monte Carlo algorithm for effciently sampling from continuous probability distributions that can take advantage of hundreds of cores and shares information between parallel Markov chains to build a scale-location mixture of Gaussians approximation to the density function of the target distribution. Expand
Using Parallel Computation to Improve Independent Metropolis–Hastings Based Estimation
Improvements to the independent Metropolis–Hastings algorithm that significantly decrease the variance of any estimator derived from the MCMC output, at a null computing cost since those improvements are based on a fixed number of target density evaluations that can be produced in parallel. Expand
Parallel multivariate slice sampling
It is demonstrated that it is possible to construct a multivariate slice sampler that has good mixing properties and is efficient in terms of computing time, and how parallel computing can be useful for making MCMC algorithms computationally efficient. Expand
Fast Hamiltonian Monte Carlo Using GPU Computing
The proposed computational framework may appear conceptually very simple, but has the potential to be applied to a wide class of hierarchical models relying on HMC sampling, whose posterior density and corresponding gradients are amenable to significant speed ups through the use of GPUs. Expand
Component-Wise Markov Chain Monte Carlo: Uniform and Geometric Ergodicity under Mixing and Composition
It is common practice in Markov chain Monte Carlo to update the simulation one variable (or sub-block of variables) at a time, rather than conduct a single full-dimensional update. When it isExpand
GPU-Accelerated Bayesian Learning and Forecasting in Simultaneous Graphical Dynamic Linear Models
A new framework of simultaneous graphical dynamic models allows the decoupling of analyses into those of a parallel set of univariate time series dynamic models, while flexibly modeling time-varying, cross-series dependencies and volatilities. Expand
Many-core algorithms for statistical phylogenetics
Novel algorithms and methods for evaluating phylogenies under arbitrary molecular evolutionary models on graphics processing units (GPUs), making use of the large number of processing cores to efficiently parallelize calculations even for large state-size models are described. Expand
Understanding GPU Programming for Statistical Computation: Studies in Massively Parallel Massive Mixtures
This article outlines important strategies and processes for GPU computation in Bayesian simulation and optimization approaches, gives examples of the benefits of GPU implementations in terms of processing speed and scale-up in ability to analyze large datasets, and provides a detailed, tutorial-style exposition that will benefit readers interested in developing GPU-based approaches in other statistical models. Expand
Bayesian Data Analysis
Detailed notes on Bayesian Computation Basics of Markov Chain Simulation, Regression Models, and Asymptotic Theorems are provided. Expand
Stan: A Probabilistic Programming Language
Stan is a probabilistic programming language for specifying statistical models that provides full Bayesian inference for continuous-variable models through Markov chain Monte Carlo methods such as the No-U-Turn sampler and an adaptive form of Hamiltonian Monte Carlo sampling. Expand