Corpus ID: 229153938

On the Treatment of Optimization Problems with L1 Penalty Terms via Multiobjective Continuation

@article{Bieker2020OnTT,
  title={On the Treatment of Optimization Problems with L1 Penalty Terms via Multiobjective Continuation},
  author={K. Bieker and Bennet Gebken and Sebastian Peitz},
  journal={ArXiv},
  year={2020},
  volume={abs/2012.07483}
}
We present a novel algorithm that allows us to gain detailed insight into the effects of sparsity in linear and nonlinear optimization, which is of great importance in many scientific areas such as image and signal processing, medical imaging, compressed sensing, and machine learning (e.g., for the training of neural networks). Sparsity is an important feature to ensure robustness against noisy data, but also to find models that are interpretable and easy to analyze due to the small number of… Expand

Figures from this paper

References

SHOWING 1-10 OF 38 REFERENCES
MOEA/D with Iterative Thresholding Algorithm for Sparse Optimization Problems
  • 16
Homotopy continuation for sparse signal representation
  • D. Malioutov, M. Çetin, A. Willsky
  • Mathematics, Computer Science
  • Proceedings. (ICASSP '05). IEEE International Conference on Acoustics, Speech, and Signal Processing, 2005.
  • 2005
  • 268
  • PDF
Fast Solution of -Norm Minimization Problems When the Solution May Be Sparse
  • 70
  • PDF
Adam: A Method for Stochastic Optimization
  • 61,126
  • PDF
Fast Solution of $\ell _{1}$ -Norm Minimization Problems When the Solution May Be Sparse
  • D. Donoho, Y. Tsaig
  • Mathematics, Computer Science
  • IEEE Transactions on Information Theory
  • 2008
  • 841
  • PDF
Constrained sparse Galerkin regression
  • 112
  • PDF
The homotopy method revisited: Computing solution paths of ℓ1-regularized problems
  • 3
  • PDF
A Survey of L 1 Regression
  • 62
  • PDF
Exploring the trade-off between generalization and empirical errors in a one-norm SVM
  • 11
L1‐regularization path algorithm for generalized linear models
  • 564
  • PDF
...
1
2
3
4
...