Information theoretic limits for linear prediction with graph-structured sparsity

@article{Barik2017InformationTL,
  title={Information theoretic limits for linear prediction with graph-structured sparsity},
  author={Adarsh Barik and Jean Honorio and Mohit Tawarmalani},
  journal={2017 IEEE International Symposium on Information Theory (ISIT)},
  year={2017},
  pages={2348-2352}
}
We analyze the necessary number of samples for sparse vector recovery in a noisy linear prediction setup. This model includes problems such as linear regression and classification. We focus on structured graph models. In particular, we prove that sufficient number of samples for the weighted graph model proposed by Hegde and others [2] is also necessary. We use the Fano's inequality [11] on well constructed ensembles as our main tool in establishing information theoretic lower bounds. 

Figures from this paper

References

SHOWING 1-10 OF 20 REFERENCES
A Nearly-Linear Time Framework for Graph-Structured Sparsity
TLDR
A framework for sparsity structures defined via graphs that is flexible and generalizes several previously studied sparsity models is introduced and achieves an information-theoretically optimal sample complexity for a wide range of parameters.
Information-theoretic bounds on model selection for Gaussian Markov random fields
TLDR
The first result establishes a set of necessary conditions on n(p, d) for any recovery method to consistently estimate the underlying graph, and the second result provides necessary conditions for any decoder to produce an estimate of the true inverse covariance matrix T satisfying ‖ Θ̂-Θ ‖ < δin the elementwise ℓ∞-norm.
Information-Theoretic Limits of Selecting Binary Graphical Models in High Dimensions
TLDR
The information-theoretic limitations of the problem of graph selection for binary Markov random fields under high-dimensional scaling, in which the graph size and the number of edges k, and/or the maximal node degree d, are allowed to increase to infinity as a function of the sample size n, are analyzed.
A fast approximation algorithm for tree-sparse recovery
TLDR
This work proposes an alternative approach to tree-sparse recovery that is based on a specific approximation algorithm for tree-projection and provably has a near-linear runtime of O(n log(kr)) and a memory cost ofO(n), where r is the dynamic range of the signal.
A Simple Proof of the Restricted Isometry Property for Random Matrices
Abstract We give a simple technique for verifying the Restricted Isometry Property (as introduced by Candès and Tao) for random matrices that underlies Compressed Sensing. Our approach has two main
Model-Based Compressive Sensing
TLDR
A model-based CS theory is introduced that parallels the conventional theory and provides concrete guidelines on how to create model- based recovery algorithms with provable performance guarantees and a new class of structured compressible signals along with a new sufficient condition for robust structured compressable signal recovery that is the natural counterpart to the restricted isometry property of conventional CS.
Sample complexity for 1-bit compressed sensing and sparse classification
This paper considers the problem of identifying the support set of a high-dimensional sparse vector, from noise-corrupted 1-bit measurements. We present passive and adaptive algorithms for this
One-Bit Compressed Sensing: Provable Support and Vector Recovery
TLDR
This paper proposes two novel and efficient solutions based on two combinatorial structures: union free families of sets and expanders for support recovery and the first method to recover a sparse vector using a near optimal number of measurements.
Subspace Pursuit for Compressive Sensing: Closing the Gap Between Performance and Complexity
TLDR
The presented analysis shows that in the noiseless setting, the proposed algorithm can exactly reconstruct arbitrary sparse signals provided that the sensing matrix satisfies the restricted isometry property with a constant parameter.
...
...