• Corpus ID: 14971610

Multivariate Regression with Grossly Corrupted Observations: A Robust Approach and its Applications

  title={Multivariate Regression with Grossly Corrupted Observations: A Robust Approach and its Applications},
  author={Xiaowei Zhang and Chi Xu and Yu Zhang and Tingshao Zhu and Li Cheng},
This paper studies the problem of multivariate linear regression where a portion of the observations is grossly corrupted or is missing, and the magnitudes and locations of such occurrences are unknown in priori. To deal with this problem, we propose a new approach by explicitly consider the error source as well as its sparseness nature. An interesting property of our approach lies in its ability of allowing individual regression output elements or tasks to possess their unique noise levels… 

Figures and Tables from this paper

A General Family of Trimmed Estimators for Robust High-dimensional Data Analysis

This work focuses on trimmed versions of structurally regularized M-estimators in the high-dimensional setting, including the popular Least Trimmed Squares estimator, as well as analogous estimators for generalized linear models and graphical models, using possibly non-convex loss functions.

DeepPrior++: Improving Fast and Accurate 3D Hand Pose Estimation

With simple improvements: adding ResNet layers, data augmentation, and better initial hand localization, DeepPrior achieves better or similar performance than more sophisticated recent methods on the three main benchmarks (NYU, ICVL, MSRA) while keeping the simplicity of the original method.

A CNN model for real time hand pose estimation

A block symmetric Gauss–Seidel decomposition theorem for convex composite quadratic programming and its applications

For a symmetric positive semidefinite linear system of equations $$\mathcal{Q}{{\varvec{x}}}= {{\varvec{b}}}$$Qx=b, where $${{\varvec{x}}}= (x_1,\ldots ,x_s)$$x=(x1,…,xs) is partitioned into s



Robust Multivariate Regression with Grossly Corrupted Observations and Its Application to Personality Prediction

This work considers the problem of multivariate linear regression with a small fraction of the responses being missing and grossly corrupted, and proposes a new algorithm that is theoretically shown to always converge to the optimal solution of its induced non-smooth optimization problem.

Outlier-Robust PCA: The High-Dimensional Case

This work proposes a high-dimensional robust principal component analysis algorithm that is efficient, robust to contaminated points, and easily kernelizable, and achieves maximal robustness.

Robust Regression via Hard Thresholding

A simple hard-thresholding algorithm called TORRENT is studied which, under mild conditions on X, can recover w* exactly even if b corrupts the response variables in an adversarial manner, i.e. both the support and entries of b are selected adversarially after observing X and w*.

Multivariate Regression with Calibration

An efficient smoothed proximal gradient algorithm which has a worst-case iteration complexity O(1/ε), where ε is a pre-specified numerical accuracy and it is proved that CMR achieves the optimal rate of convergence in parameter estimation.

Smoothing proximal gradient method for general structured sparse regression

This paper proposes a general optimization approach, the smoothing proximal gradient method, which can solve structured sparse regression problems with any smooth convex loss under a wide spectrum of structured sparsity-inducing penalties.

Estimation of high-dimensional low-rank matrices

This work investigates penalized least squares estimators with a Schatten-p quasi-norm penalty term and derives bounds for the kth entropy numbers of the quasi-convex Schatten class embeddings S M p → S M 2 , p < 1, which are of independent interest.

Structured Sparsity via Alternating Direction Methods

This paper proposes a unified framework based on the augmented Lagrangian method, under which problems with both types of regularization and their variants can be efficiently solved, and develops new algorithms using an alternating partial-linearization/splitting technique.

Dense error correction via l1-minimization

It is proved that for highly correlated dictionaries A, any non-negative, sufficiently sparse signal x can be recovered by solving an l1-minimization problem: min ‖x‖ 1 + ‖e‚ 1 subject to y = Ax + e, which suggests that accurate and efficient recovery of sparse signals is possible even with nearly 100% of the observations corrupted.

Convex multi-task feature learning

It is proved that the method for learning sparse representations shared across multiple tasks is equivalent to solving a convex optimization problem for which there is an iterative algorithm which converges to an optimal solution.

Oracle Inequalities and Optimal Inference under Group Sparsity

The Group Lasso can achieve an improvement in the prediction and estimation properties as compared to the Lasso, and it is proved that the rate of convergence of the upper bounds is optimal in a minimax sense.