• Corpus ID: 88522121

Variable Selection for High-dimensional Generalized Linear Models using an Iterated Conditional Modes/Medians Algorithm

@article{Pungpapong2017VariableSF,
  title={Variable Selection for High-dimensional Generalized Linear Models using an Iterated Conditional Modes/Medians Algorithm},
  author={Vitara Pungpapong and Min Zhang and Dabao Zhang},
  journal={arXiv: Methodology},
  year={2017}
}
High-dimensional linear and nonlinear models have been extensively used to identify associations between response and explanatory variables. The variable selection problem is commonly of interest in the presence of massive and complex data. An empirical Bayes model for high-dimensional generalized linear models (GLMs) is considered in this paper. The extension of the Iterated Conditional Modes/Medians (ICM/M) algorithm is proposed to build up a GLM. With the construction of pseudodata and… 
Greedy forward regression for variable screening
TLDR
A new and simple screening method that incorporates multiple predictors at each step of forward regression, with decisions on which variables to incorporate based on the same criterion regarded as a generalisation and unification of FR and SIS.

References

SHOWING 1-10 OF 35 REFERENCES
Selecting massive variables using an iterated conditional modes/medians algorithm
TLDR
An iterated conditional modes/medians (ICM/M) algorithm is proposed to implement empirical Bayes selection of massive variables, while incorporating sparsity or more complicated a priori information.
Spike and slab variable selection: Frequentist and Bayesian strategies
TLDR
This paper introduces a variable selection method referred to as a rescaled spike and slab model, and studies the usefulness of continuous bimodal priors to model hypervariance parameters, and the effect scaling has on the posterior mean through its relationship to penalization.
EMVS: The EM Approach to Bayesian Variable Selection
TLDR
EMVS is proposed, a deterministic alternative to stochastic search based on an EM algorithm which exploits a conjugate mixture prior formulation to quickly find posterior modes in high-dimensional linear regression contexts.
The group lasso for logistic regression
TLDR
An efficient algorithm is presented, that is especially suitable for high dimensional problems, which can also be applied to generalized linear models to solve the corresponding convex optimization problem.
Adaptive Lasso for sparse high-dimensional regression models
TLDR
The adaptive Lasso has the oracle property even when the number of covariates is much larger than the sample size, and under a partial orthogonality condition in which the covariates with zero coefficients are weakly correlated with the covariate with nonzero coefficients, marginal regression can be used to obtain the initial estimator.
APPROACHES FOR BAYESIAN VARIABLE SELECTION
This paper describes and compares various hierarchical mixture prior formulations of variable selection uncertainty in normal linear regression models. These include the nonconjugate SSVS formulation
Partial least squares Cox regression for genome-wide data
TLDR
This work proposes a modification of the method of park et al. (2002) such that estimates of the baseline hazard and the gene effects are obtained in separate steps, thereby enabling estimation by iteratively reweighted partial least squares for generalized linear models.
Detecting differential gene expression with a semiparametric hierarchical mixture method.
TLDR
This work proposes a hierarchical mixture model to provide methodology that is both sensitive in detecting differential expression and sufficiently flexible to account for the complex variability of normalized microarray data.
L1 Regularization Path Algorithm for Generalized Linear Models
TLDR
An estimation algorithm of coefficient to select variables for L1 regularized generalized linear models to efficiently compute solutions along the entire regularization path using the predictor-corrector method of convex-optimization.
Regression Shrinkage and Selection via the Lasso
TLDR
A new method for estimation in linear models called the lasso, which minimizes the residual sum of squares subject to the sum of the absolute value of the coefficients being less than a constant, is proposed.
...
...