Bregman Proximal Framework for Deep Linear Neural Networks
@article{Mukkamala2019BregmanPF, title={Bregman Proximal Framework for Deep Linear Neural Networks}, author={Mahesh Chandra Mukkamala and Felix Westerkamp and E. Laude and D. Cremers and P. Ochs}, journal={ArXiv}, year={2019}, volume={abs/1910.03638} }
A typical assumption for the analysis of first order optimization methods is the Lipschitz continuity of the gradient of the objective function. However, for many practical applications this assumption is violated, including loss functions in deep learning. To overcome this issue, certain extensions based on generalized proximity measures known as Bregman distances were introduced. This initiated the development of the Bregman proximal gradient (BPG) algorithm and an inertial variant (momentum… CONTINUE READING
3 Citations
Global Convergence of Model Function Based Bregman Proximal Minimization Algorithms
- Computer Science, Mathematics
- ArXiv
- 2020
- PDF
First-Order Algorithms Without Lipschitz Gradient: A Sequential Local Optimization Approach
- Mathematics
- 2020
- PDF
Convex-Concave Backtracking for Inertial Bregman Proximal Gradient Algorithms in Non-Convex Optimization
- Computer Science, Mathematics
- SIAM J. Math. Data Sci.
- 2020
- 12
- PDF
References
SHOWING 1-10 OF 37 REFERENCES
Beyond Alternating Updates for Matrix Factorization with Inertial Bregman Proximal Gradient Algorithms
- Computer Science, Mathematics
- NeurIPS
- 2019
- 6
- PDF
Bregman Proximal Gradient Algorithm With Extrapolation for a Class of Nonconvex Nonsmooth Minimization Problems
- Mathematics, Computer Science
- IEEE Access
- 2019
- 3
- PDF
Provable Bregman-divergence based Methods for Nonconvex and Non-Lipschitz Problems
- Computer Science, Mathematics
- ArXiv
- 2019
- 14
- PDF
First Order Methods beyond Convexity and Lipschitz Gradient Continuity with Applications to Quadratic Inverse Problems
- Computer Science, Mathematics
- SIAM J. Optim.
- 2018
- 60
- PDF
A Descent Lemma Beyond Lipschitz Gradient Continuity: First-Order Methods Revisited and Applications
- Mathematics, Computer Science
- Math. Oper. Res.
- 2017
- 150
- Highly Influential
- PDF
Adaptive Subgradient Methods for Online Learning and Stochastic Optimization
- Computer Science, Mathematics
- J. Mach. Learn. Res.
- 2011
- 6,536
- PDF
Implicit Regularization of Discrete Gradient Dynamics in Deep Linear Neural Networks
- Computer Science, Mathematics
- NeurIPS
- 2019
- 34
- PDF
Inertial Block Mirror Descent Method for Non-Convex Non-Smooth Optimization
- Mathematics, Computer Science
- 2019
- 7
- PDF
A Block Coordinate Descent Method for Regularized Multiconvex Optimization with Applications to Nonnegative Tensor Factorization and Completion
- Mathematics, Computer Science
- SIAM J. Imaging Sci.
- 2013
- 686
- PDF