Bayesian Model Selection and Model Averaging.

@article{Wasserman2000BayesianMS,
  title={Bayesian Model Selection and Model Averaging.},
  author={Wasserman},
  journal={Journal of mathematical psychology},
  year={2000},
  volume={44 1},
  pages={
          92-107
        }
}
  • Wasserman
  • Published 1 March 2000
  • Economics
  • Journal of mathematical psychology
This paper reviews the Bayesian approach to model selection and model averaging. In this review, I emphasize objective Bayesian methods based on noninformative priors. I will also discuss implementation details, approximations, and relationships to other methods. Copyright 2000 Academic Press. 

Tables from this paper

On model selection in cosmology
We review some of the common methods for model selection: the goodness of fit, the likelihood ratio test, Bayesian model selection using Bayes factors, and the classical as well as the Bayesian
Using the Bayesian information criterion to develop two-stage model-robust and model-sensitive designs
TLDR
Use of the Bayesian Information Criterion (BIC) is investigated in the development of Bayesian two-stage designs robust to model uncertainty and can readily be computed from the output of standard statistical software packages.
Bayesian model comparison via parallel model output
Bayesian estimation via MCMC methods opens up new possibilities in estimating complex models. However, there is still considerable debate about how selection among a set of candidate models, or
Data-Based Priors for Bayesian Model Averaging
TLDR
Simulations show that by using proper priors for Bayesian model averaging, smaller mean squared prediction errors can be gotten both in synthetic data and real data studies, especially for the data of poor quality.
Normative selection of Bayesian networks
Bayesian model choice based on Monte Carlo estimates of posterior model probabilities
Bayesian Model Averaging in Multi-Factor Markets
The Bayesian mechanism can be applied to update information on return forecasts derived from multi-factor models. An out-of-sample backtest reveals distinct features of Bayesian model averaging for
Loss Distribution Approach
This chapter introduces a basic model for the Loss Distribution Approach. We discuss the main aspects of the model and basic probabilistic concepts of risk quantification. The essentials of the
Bayesian model selection and averaging via mixture model estimation
TLDR
A new approach for Bayesian model averaging and selection is proposed, based on the mixture model approach for hypothesis testing, that shows that this posterior distribution is equal to the ‘Bayesian-model averaged’ posterior distribution over all candidate models, weighted by their posterior probability.
Bayesian inference of engineering models
TLDR
A new method for numerical Bayesian inference, termed aBUS, is proposed, which is based on Subset Simulation applied within BUS, and a Bayesian framework for probabilistic modeling is discussed based on Cox-Jaynes axioms of probability theory.
...
...

References

SHOWING 1-10 OF 28 REFERENCES
The Intrinsic Bayes Factor for Model Selection and Prediction
TLDR
This article introduces a new criterion called the intrinsic Bayes factor, which is fully automatic in the sense of requiring only standard noninformative priors for its computation and yet seems to correspond to very reasonable actual Bayes factors.
BAYES AND EMPIRICAL BAYES METHODS FOR DATA ANALYSIS
TLDR
Approaches for Statistical Inference: The Bayes Approach, Model Criticism and Selection, and Performance of Bayes Procedures.
Sampling-Based Approaches to Calculating Marginal Densities
Abstract Stochastic substitution, the Gibbs sampler, and the sampling-importance-resampling algorithm can be viewed as three alternative sampling- (or Monte Carlo-) based approaches to the
Information Theory and an Extension of the Maximum Likelihood Principle
TLDR
The classical maximum likelihood principle can be considered to be a method of asymptotic realization of an optimum estimate with respect to a very general information theoretic criterion to provide answers to many practical problems of statistical model fitting.
Fractional Bayes factors for model comparison
Bayesian comparison of models is achieved simply by calculation of posterior probabilities of the models themselves. However, there are difficulties with this approach when prior information about
Bayesian Model Selection in Social Research
It is argued that P-values and the tests based upon them give unsatisfactory results, especially in large samples. It is shown that, in regression, when there are many candidate independent
Alternative Bayes factors for model selection
Several alternative Bayes factors have been recently proposed in order to solve the problem of the extreme sensitivity of the Bayes factor to the priors of models under comparison. Specifically, the
Posterior Predictive Assessment of Model Fitnessvia Realized
This paper considers the Bayesian counterparts of the classical tests for goodness of t and their use in judging the t of a single Bayesian model to the observed data. We focus on posterior
A Reference Bayesian Test for Nested Hypotheses and its Relationship to the Schwarz Criterion
Abstract To compute a Bayes factor for testing H 0: ψ = ψ0 in the presence of a nuisance parameter β, priors under the null and alternative hypotheses must be chosen. As in Bayesian estimation, an
POSTERIOR PREDICTIVE ASSESSMENT OF MODEL FITNESS VIA REALIZED DISCREPANCIES
This paper considers Bayesian counterparts of the classical tests for good- ness of fit and their use in judging the fit of a single Bayesian model to the observed data. We focus on posterior
...
...