• Corpus ID: 237513911

# How to use KL-divergence to construct conjugate priors, with well-defined non-informative limits, for the multivariate Gaussian

```@inproceedings{Brummer2021HowTU,
title={How to use KL-divergence to construct conjugate priors, with well-defined non-informative limits, for the multivariate Gaussian},
author={Niko Brummer},
year={2021}
}```
• N. Brummer
• Published 15 September 2021
• Mathematics, Computer Science
The Wishart distribution is the standard conjugate prior for the precision of the multivariate Gaussian likelihood, when the mean is known—while the normal-Wishart can be used when the mean is also unknown. It is however not so obvious how to assign values to the hyperparameters of these distributions. In particular, when forming non-informative limits of these distributions, the shape (or degrees of freedom) parameter of the Wishart must be handled with care. The intuitive solution of directly…

## References

SHOWING 1-6 OF 6 REFERENCES
Inferring a Gaussian distribution
A common question in statistical modeling is “which out of a continuum of models are likely to have generated this data?” For the Gaussian class of models, this question can be answered completely
Some matrix-variate distribution theory: Notational considerations and a Bayesian application
SUMMARY We introduce and justify a convenient notation for certain matrix-variate distributions which, by its emphasis on the important underlying parameters, and the theory on which it is based,
Probability theory: the logic of science
This is a remarkable book by a remarkable scientist. E. T. Jaynes was a physicist, principally theoretical, who found himself driven to spend much of his life advocating, defending and developing a
Pattern Recognition and Machine Learning
This book covers a broad range of topics for regular factorial designs and presents all of the material in very mathematical fashion and will surely become an invaluable resource for researchers and graduate students doing research in the design of factorial experiments.
Murhpy, Machine Learning: a Probabilistic Perspective
• 2012
The term − d 2 log α was absorbed into const in line 4