# On the Convergence of Block Coordinate Descent Type Methods

@article{Beck2013OnTC,
title={On the Convergence of Block Coordinate Descent Type Methods},
author={Amir Beck and Luba Tetruashvili},
journal={SIAM J. Optim.},
year={2013},
volume={23},
pages={2037-2060}
}
• Published 16 October 2013
• Mathematics, Computer Science
• SIAM J. Optim.
In this paper we study smooth convex programming problems where the decision variables vector is split into several blocks of variables. We analyze the block coordinate gradient projection method in which each iteration consists of performing a gradient projection step with respect to a certain block taken in a cyclic order. Global sublinear rate of convergence of this method is established and it is shown that it can be accelerated when the problem is unconstrained. In the unconstrained…
422 Citations
The Cyclic Block Conditional Gradient Method for Convex Optimization Problems
• Computer Science, Mathematics
SIAM J. Optim.
• 2015
Numerical comparisons of the proposed method to both the classical conditional gradient algorithm and its random block version demonstrate the effectiveness of the cyclic block update rule.
On the Convergence of a Regularized Jacobi Algorithm for Convex Optimization
• Mathematics, Computer Science
IEEE Transactions on Automatic Control
• 2018
The convergence analysis of the regularized Jacobi algorithm is revisited and it is shown that it also converges in iterates under very mild conditions on the objective function and achieves a linear convergence rate.
Parallel coordinate descent methods for composite minimization: convergence analysis and error bounds
• Mathematics, Computer Science
• 2013
It is shown that the theoretical estimates on the convergence rate depend on the number of blocks chosen randomly and a natural measure of separability of the objective function, and it is proved that the new class of generalized error bound functions encompasses both global/local error bound function and smooth strongly convex functions.
On the convergence of a Block-Coordinate Incremental Gradient method
• Mathematics
Soft Comput.
• 2021
It is proved that the block-coordinate incremental gradient method can be seen as a gradient method with errors and convergence can be proved by showing the error at each iteration satisfies some standard conditions.
Iteration complexity analysis of block coordinate descent methods
• Computer Science, Mathematics
Math. Program.
• 2017
This paper unify these algorithms under the so-called block successive upper-bound minimization (BSUM) framework, and shows that for a broad class of multi-block nonsmooth convex problems, all algorithms achieve a global sublinear iteration complexity of O(1/r), where r is the iteration index.
The 2-Coordinate Descent Method for Solving Double-Sided Simplex Constrained Minimization Problems
• A. Beck
• Mathematics
J. Optim. Theory Appl.
• 2014
This paper considers the problem of minimizing a continuously differentiable function with a Lipschitz continuous gradient subject to a single linear equality constraint and additional bound
The Analysis of Alternating Minimization Method for Double Sparsity Constrained Optimization Problem
• Mathematics, Computer Science
Asia Pac. J. Oper. Res.
• 2020
This work analyzes the alternating minimization (AM) method for solving double sparsity constrained minimization problem, where the decision variable vector is split into two blocks and establishes a non-asymptotic sub-linear rate of convergence under the assumption of convexity.
Parallel Random Coordinate Descent Method for Composite Minimization: Convergence Analysis and Error Bounds
• Mathematics, Computer Science
SIAM J. Optim.
• 2016
A parallel version of a randomized (block) coordinate descent method for minimizing the sum of a partially separable smooth convex function and a fully separable nonsmooth convexfunction has a sublinear convergence rate.
A block coordinate descent method of multipliers: Convergence analysis and applications
• Mathematics, Computer Science
2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)
• 2014
This work proposes a new class of algorithms called the block coordinate descent method of multipliers (BCDMM) to solve a nonsmooth convex problem with linear coupling constraints and shows that under certain regularity conditions, the BCDMM converges to the set of optimal solutions.
A proximal block minimization method of multipliers with a substitution procedure
• Computer Science, Mathematics
Optim. Methods Softw.
• 2015
This paper proposes a new algorithm called the proximal block minimization method of multipliers with a substitution to solve this family of problems of this form and proves its convergence via the analytic framework of contractive type methods and derives a worst-case convergence rate in an ergodic sense.