# Variable Selection for High-dimensional Generalized Linear Models using an Iterated Conditional Modes/Medians Algorithm

@article{Pungpapong2017VariableSF, title={Variable Selection for High-dimensional Generalized Linear Models using an Iterated Conditional Modes/Medians Algorithm}, author={Vitara Pungpapong and Min Zhang and Dabao Zhang}, journal={arXiv: Methodology}, year={2017} }

High-dimensional linear and nonlinear models have been extensively used to identify associations between response and explanatory variables. The variable selection problem is commonly of interest in the presence of massive and complex data. An empirical Bayes model for high-dimensional generalized linear models (GLMs) is considered in this paper. The extension of the Iterated Conditional Modes/Medians (ICM/M) algorithm is proposed to build up a GLM. With the construction of pseudodata and…

## Figures and Tables from this paper

## One Citation

Greedy forward regression for variable screening

- Computer Science
- 2015

A new and simple screening method that incorporates multiple predictors at each step of forward regression, with decisions on which variables to incorporate based on the same criterion regarded as a generalisation and unification of FR and SIS.

## References

SHOWING 1-10 OF 35 REFERENCES

Selecting massive variables using an iterated conditional modes/medians algorithm

- Computer Science
- 2015

An iterated conditional modes/medians (ICM/M) algorithm is proposed to implement empirical Bayes selection of massive variables, while incorporating sparsity or more complicated a priori information.

Spike and slab variable selection: Frequentist and Bayesian strategies

- Computer Science
- 2005

This paper introduces a variable selection method referred to as a rescaled spike and slab model, and studies the usefulness of continuous bimodal priors to model hypervariance parameters, and the effect scaling has on the posterior mean through its relationship to penalization.

EMVS: The EM Approach to Bayesian Variable Selection

- Computer Science
- 2014

EMVS is proposed, a deterministic alternative to stochastic search based on an EM algorithm which exploits a conjugate mixture prior formulation to quickly find posterior modes in high-dimensional linear regression contexts.

The group lasso for logistic regression

- Computer Science
- 2008

An efficient algorithm is presented, that is especially suitable for high dimensional problems, which can also be applied to generalized linear models to solve the corresponding convex optimization problem.

Adaptive Lasso for sparse high-dimensional regression models

- Mathematics, Computer Science
- 2008

The adaptive Lasso has the oracle property even when the number of covariates is much larger than the sample size, and under a partial orthogonality condition in which the covariates with zero coefficients are weakly correlated with the covariate with nonzero coefficients, marginal regression can be used to obtain the initial estimator.

APPROACHES FOR BAYESIAN VARIABLE SELECTION

- Mathematics
- 1997

This paper describes and compares various hierarchical mixture prior formulations of variable selection uncertainty in normal linear regression models. These include the nonconjugate SSVS formulation…

Partial least squares Cox regression for genome-wide data

- MathematicsLifetime data analysis
- 2008

This work proposes a modification of the method of park et al. (2002) such that estimates of the baseline hazard and the gene effects are obtained in separate steps, thereby enabling estimation by iteratively reweighted partial least squares for generalized linear models.

Detecting differential gene expression with a semiparametric hierarchical mixture method.

- BiologyBiostatistics
- 2004

This work proposes a hierarchical mixture model to provide methodology that is both sensitive in detecting differential expression and sufficiently flexible to account for the complex variability of normalized microarray data.

L1 Regularization Path Algorithm for Generalized Linear Models

- Computer Science
- 2008

An estimation algorithm of coefficient to select variables for L1 regularized generalized linear models to efficiently compute solutions along the entire regularization path using the predictor-corrector method of convex-optimization.

Regression Shrinkage and Selection via the Lasso

- Computer Science
- 1996

A new method for estimation in linear models called the lasso, which minimizes the residual sum of squares subject to the sum of the absolute value of the coefficients being less than a constant, is proposed.