• Corpus ID: 250264457

# Maximum a posteriori estimators in $\ell^p$ are well-defined for diagonal Gaussian priors

@inproceedings{Klebanov2022MaximumAP,
title={Maximum a posteriori estimators in \$\ell^p\$ are well-defined for diagonal Gaussian priors},
author={Ilja Klebanov and Philipp Wacker},
year={2022}
}
• Published 1 July 2022
• Mathematics
We prove that maximum a posteriori estimators are well-deﬁned for diagonal Gaussian priors µ on (cid:96) p under common assumptions on the potential Φ . Further, we show connections to the Onsager–Machlup functional and provide a corrected and strongly simpliﬁed proof in the Hilbert space case p = 2, previously established by Dashti et al. (2013); Kretschmann (2019).
1 Citations

## Figures from this paper

### An order-theoretic perspective on modes and maximum a posteriori estimation in Bayesian inverse problems

• Mathematics
• 2022
. It is often desirable to summarise a probability measure on a space X in terms of a mode, or MAP estimator, i.e. a point of maximum probability. Such points can be rigorously deﬁned using masses of

## References

SHOWING 1-10 OF 21 REFERENCES

### Sparsity-promoting and edge-preserving maximum a posteriori estimators in non-parametric Bayesian inverse problems

• Mathematics, Computer Science
• 2017
This work considers the inverse problem of recovering an unknown functional parameter u in a separable Banach space, from a noisy observation vector y of its image through a known possibly non-linear map G, and shows that in this non-parametric setup the maximum a posteriori (MAP) estimates are characterized by the minimizers of a generalized Onsager–Machlup functional of the posterior.

### MAP estimators and their consistency in Bayesian nonparametric inverse problems

• Mathematics
• 2013
We consider the inverse problem of estimating an unknown function u from noisy measurements y of a known, possibly nonlinear, map G?> applied to u. We adopt a Bayesian approach to the problem and

### Maximum a posteriori probability estimates in infinite-dimensional Bayesian inverse problems

• Mathematics
• 2014
A demanding challenge in Bayesian inversion is to efficiently characterize the posterior distribution. This task is problematic especially in high-dimensional non-Gaussian problems, where the

### Γ -convergence of Onsager–Machlup functionals: I. With applications to maximum a posteriori estimation in Bayesian inverse problems

• Mathematics
Inverse Problems
• 2021
A general convergence theory for modes that is based on the Γ-convergence of OM functionals is proposed and applied to Bayesian inverse problems with Gaussian and edge-preserving Besov priors.

### Nonparametric Bayesian inverse problems with Laplacian noise

The focus of this work are Bayesian inverse problems in an infinite-dimensional setting with Gaussian prior and data corrupted by additive Laplacian noise. In particular, the connection between

### Generalized Modes in Bayesian Inverse Problems

• Mathematics
SIAM/ASA J. Uncertain. Quantification
• 2019
Uncertainty quantification requires efficient summarization of high- or even infinite-dimensional distributions based on, e.g., suitable point estimates (modes) for posterior ...

### Γ-convergence of Onsager–Machlup functionals: II. Infinite product measures on Banach spaces

• Mathematics
Inverse Problems
• 2021
This work derives Onsager–Machlup functionals for countable product measures on weighted ℓ p subspaces of the sequence space RN that provides a basis for analysis of the convergence of maximum a posteriori estimators in Bayesian inverse problems and most likely paths in transition path theory.

### Equivalence of weak and strong modes of measures on topological vector spaces

• Mathematics
Inverse Problems
• 2018
A strong mode of a probability measure on a normed space X can be defined as a point u such that the mass of the ball centred at u uniformly dominates the mass of all other balls in the small-radius

### Joint measures and cross-covariance operators

Let H1 (resp., H2) be a real and separable Hilbert space with Borel o-field r1 (resp., r2), and let (H1 x H2, r, x r2) be the product measurable space generated by the measurable rectangles. This

### Inverse problems: A Bayesian perspective

The Bayesian approach to regularization is reviewed, developing a function space viewpoint on the subject, which allows for a full characterization of all possible solutions, and their relative probabilities, whilst simultaneously forcing significant modelling issues to be addressed in a clear and precise fashion.