A gradient estimator via L1-randomization for online zero-order optimization with two point feedback
@inproceedings{Akhavan2022AGE, title={A gradient estimator via L1-randomization for online zero-order optimization with two point feedback}, author={Arya Akhavan and Evgenii Chzhen and Massimiliano Pontil and A. Tsybakov}, year={2022} }
This work studies online zero-order optimization of convex and Lipschitz functions. We present a novel gradient estimator based on two function evaluations and randomization on the (cid:96) 1 -sphere. Considering different geometries of feasible sets and Lipschitz assumptions we analyse online dual averaging algorithm with our estimator in place of the usual gradient. We consider two types of assumptions on the noise of the zero-order oracle: canceling noise and adversarial noise. We provide an…
Figures from this paper
3 Citations
The power of first-order smooth optimization for black-box non-smooth problems
- Computer ScienceICML
- 2022
A generic approach is proposed that, based on optimal first-order methods, allows to obtain in a black-box fashion new zeroth-order algorithms for non-smooth convex optimization problems and elaborate on extensions for stochastic optimization problems, saddle-point problems, and distributed optimization.
Randomized gradient-free methods in convex optimization
- Computer Science
- 2022
This review presents modern gradient-free methods to solve convex optimization problems and mainly focuses on three criteria: oracle complexity, iteration complexity, and the maximum permissible noise level.
Gradient-Free Federated Learning Methods with $l_1$ and $l_2$-Randomization for Non-Smooth Convex Stochastic Optimization Problems
- Economics
- 2022
The goal of this paper is to build on the current advances in gradient-free non-smooth optimization and in feild of federated learning, gradient- free methods for solving non-Smooth stochastic optimization problems in federatedlearning architecture.
References
SHOWING 1-10 OF 33 REFERENCES
Distributed Zero-Order Optimization under Adversarial Noise
- Computer Science, MathematicsNeurIPS
- 2021
We study the problem of distributed zero-order optimization for a class of strongly convex functions. They are formed by the average of local objectives, associated to different nodes in a prescribed…
A Modern Introduction to Online Learning
- Computer ScienceArXiv
- 2019
This monograph introduces the basic concepts of Online Learning through a modern view of Online Convex Optimization, and presents first-order and second-order algorithms for online learning with convex losses, in Euclidean and non-Euclidean settings.
Exploiting Higher Order Smoothness in Derivative-free Optimization and Continuous Bandits
- Computer Science, MathematicsNeurIPS
- 2020
The results imply that the zero-order algorithm is nearly optimal in terms of sample complexity and the problem parameters, and an estimator of the minimum value of the function achieving almost sharp oracle behavior is proposed.
Highly-Smooth Zero-th Order Online Optimization
- Computer Science, MathematicsCOLT
- 2016
It is shown that as opposed to gradient-based algorithms, high-order smoothness may be used to improve estimation rates, with a precise dependence of the authors' upper-bounds on the degree of smoothness.
Kernel-based methods for bandit convex optimization
- Computer ScienceSTOC
- 2017
We consider the adversarial convex bandit problem and we build the first poly(T)-time algorithm with poly(n) √T-regret for this problem. To do so we introduce three new ideas in the derivative-free…
Improved Regret Guarantees for Online Smooth Convex Optimization with Bandit Feedback
- Computer Science, MathematicsAISTATS
- 2011
The first algorithm whose expected regret is O(T ), ignoring constant and logarithmic factors is given, building upon existing work on selfconcordant regularizers and one-point gradient estimation.
On the Complexity of Bandit and Derivative-Free Stochastic Convex Optimization
- Computer Science, MathematicsCOLT
- 2013
The attainable error/regret in the bandit and derivative-free settings, as a function of the dimension d and the available number of queries T is investigated, and a precise characterization of the attainable performance for strongly-convex and smooth functions is provided.
Optimal Rates for Zero-Order Convex Optimization: The Power of Two Function Evaluations
- Computer Science, MathematicsIEEE Transactions on Information Theory
- 2015
Focusing on nonasymptotic bounds on convergence rates, it is shown that if pairs of function values are available, algorithms for d-dimensional optimization that use gradient estimates based on random perturbations suffer a factor of at most √d in convergence rate over traditional stochastic gradient methods.
Escaping the Local Minima via Simulated Annealing: Optimization of Approximately Convex Functions
- Computer ScienceCOLT
- 2015
We consider the problem of optimizing an approximately convex function over a bounded convex set in $\mathbb{R}^n$ using only function evaluations. The problem is reduced to sampling from an…