Full adaptation to smoothness using randomly truncated series priors with Gaussian coefficients and inverse gamma scaling

@article{Waaij2017FullAT,
  title={Full adaptation to smoothness using randomly truncated series priors with Gaussian coefficients and inverse gamma scaling},
  author={Jan van Waaij and Harry van Zanten},
  journal={Statistics \& Probability Letters},
  year={2017},
  volume={123},
  pages={93-99}
}
4 Citations

Adaptive posterior contraction rates for empirical Bayesian drift estimation of a diffusion

Due to their conjugate posteriors, Gaussian process priors are attractive for estimating the drift of stochastic differential equations with continuous time observations. However, their performance

Adaptive posterior contraction rates for diffusions

Diffusions have many applications in science and can be described with a stochastic differential equation (SDE). We consider the following SDE, which was for example used in moleculair dynamics (see

Nonparametric Bayesian label prediction on a large graph using truncated Laplacian regularization

TLDR
An implementation of a nonparametric Bayesian approach to solving binary classification problems on graphs with a prior constructed by truncating a series expansion of the soft label function using the graph Laplacian eigenfunctions as basisfunctions is described.

Nonparametric Bayesian label prediction on a graph

References

SHOWING 1-10 OF 15 REFERENCES

Adaptive estimation of multivariate functions using conditionally Gaussian tensor-product spline priors

We investigate posterior contraction rates for priors on multi- variate functions that are constructed using tensor-product B-spline expan- sions. We prove that using a hierarchical prior with an

Bayesian Optimal Adaptive Estimation Using a Sieve Prior

TLDR
It is possible to prove that the adaptive Bayesian approach for the l2 loss is strongly suboptimal and the rate is provided, and a lower bound is provided on the rate.

Gaussian process methods for one-dimensional diffusions: optimal rates and adaptation

TLDR
This work improves existing convergence rate results for Gaussian process (GP) priors with fixed hyper parameters by considering hierarchical procedures that involve either a prior on a multiplicative scaling parameter, or aPrior on the regularity parameter of the GP.

Adaptive Bayesian Procedures Using Random Series Priors

We consider a general class of prior distributions for nonparametric Bayesian estimation which uses finite random series with a random number of terms. A prior is constructed through distributions on

Convergence rates of posterior distributions for non-i.i.d. observations

We consider the asymptotic behavior of posterior distributions and Bayes estimators based on observations which are required to be neither independent nor identically distributed. We give general

CONVERGENCE RATES OF POSTERIOR DISTRIBUTIONS FOR NONIID OBSERVATIONS By

We consider the asymptotic behavior of posterior distributions and Bayes estimators based on observations which are required to be neither independent nor identically distributed. We give general

Bayesian aspects of some nonparametric problems

TLDR
All Bayes estimators for proper Gaussian priors have zero asymptotic efficiency in this minimax sense, and a class of priors whose Bayes procedures attain the optimal minimax rate of convergence is presented.

Convergence rates of posterior distributions for Brownian semimartingale models

We consider the asymptotic behaviour of posterior distributions based on continuous observations from a Brownian semimartingale model. We present a general result that bounds the posterior rate of

Reversible jump MCMC for nonparametric drift estimation for diffusion processes

Reproducing kernel Hilbert spaces of Gaussian priors

We review definitions and properties of reproducing kernel Hilbert spaces attached to Gaussian variables and processes, with a view to applications in nonparametric Bayesian statistics using Gaussian