Parallel integrative learning for large-scale multi-response regression with incomplete outcomes

  title={Parallel integrative learning for large-scale multi-response regression with incomplete outcomes},
  author={Ruipeng Dong and Daoji Li and Zemin Zheng},
  journal={Comput. Stat. Data Anal.},
1 Citations

Figures and Tables from this paper



Sequential Co-Sparse Factor Regression

  • Aditya MishraD. DeyKun Chen
  • Computer Science
    Journal of computational and graphical statistics : a joint publication of American Statistical Association, Institute of Mathematical Statistics, Interface Foundation of North America
  • 2017
An efficient computational procedure, named sequential factor extraction via co-sparse unit-rank estimation (SeCURE), that completely bypasses the orthogonality requirements of multivariate regression models and ensures efficient computation even with incomplete data and/or when enforcing exact orthog onality is desired.


We show that, under a sparsity scenario, the Lasso estimator and the Dantzig selector exhibit similar behavior. For both methods, we derive, in parallel, oracle inequalities for the prediction risk

Tuning parameter selection in high dimensional penalized likelihood

It is shown that using the Akaike information criterion or Bayes information criterion to select the tuning parameter may not be adequate for consistently identifying the true model, and a uniform choice of the model complexity penalty is proposed.

Low Rank Matrix Completion with Exponential Family Noise

This work considers an estimator defined as the minimizer of the sum of a log-likelihood term and a nuclear norm penalization and proves an upper bound on the Frobenius prediction risk, and shows that all the rates obtained are minimax optimal up to a logarithmic factor.

Sparsity oracle inequalities for the Lasso

It is shown that the penalized least squares estimator satisfies sparsity oracle inequalities, i.e., bounds in terms of the number of non-zero components of the oracle vector, in nonparametric regression setting with random design.

Scalable Interpretable Multi-Response Regression via SEED

A scalable procedure called sequential estimation with eigen-decomposition (SEED) which needs only a single top-$r$ singular value decomposition to find the optimal low-rank and sparse matrix by solving a sparse generalized eigenvalue problem.

Sparse reduced-rank regression for integrating omics data

The proposed method is used to integrate genomics and metabolomics data to identify genetic variants that are potentially predictive of atherosclerosis cardiovascular disease (ASCVD) beyond well-established risk factors.

Multivariate Functional Regression Via Nested Reduced-Rank Regularization

Abstract We propose a nested reduced-rank regression (NRRR) approach in fitting a regression model with multivariate functional responses and predictors to achieve tailored dimension reduction and

Multivariate spatial autoregressive model for large scale social networks

Dimensionality Reduction and Variable Selection in Multivariate Varying-Coefficient Models With a Large Number of Covariates

ABSTRACT Motivated by the study of gene and environment interactions, we consider a multivariate response varying-coefficient model with a large number of covariates. The need of nonparametrically