Variational Analysis

@inproceedings{Rockafellar1998VariationalA,
  title={Variational Analysis},
  author={R. Tyrrell Rockafellar and Roger J.-B. Wets},
  booktitle={Grundlehren der mathematischen Wissenschaften},
  year={1998}
}
Errata and Additions: December 2013 p. 10, l. 14 xν → x should be xν → x̄ p. 31, l. 10 argmin f should be argmin f̄ p. 45, Fig.2-7 the x-axis should be just IR not IRn p. 108, l. -6 insert ‘to’ after ‘said’ p. 136, l. -11 Proposition 4.37, replace xinIB by x ∈ IB p. 278, l. 13 replace ‘neq’ by 6= p. 514, l. 19 replace φ(x) by φ(y) p. 644, l. -7 S(B) should be S(B) p. 677, l. -1 replace Sε : T 7→ IR n by Sε : T → IR n p. 678, l. 4 in the expression for Aν replace κ by ν p. 734, l. -2 replace 38… 

Figures and Topics from this paper

Alternating proximal algorithms for linearly constrained variational inequalities: Application to do
Abstract Let X , Y , Z be real Hilbert spaces, let f : X → R ∪ { + ∞ } , g : Y → R ∪ { + ∞ } be closed convex functions and let A : X → Z , B : Y → Z be linear continuous operators. Let us consider
Nonsmooth Kantorovich–Newton Methods: Hypotheses and Auxiliary Problems
In Newton’s method 0 ∈ f(xk) + G(xk)(xk+ 1 − xk) for solving a nonsmooth equation f(x) = 0, the type of approximation of f by some (generally multivalued) mapping G determines not only the
A study of Nesterov's scheme for Lagrangian decomposition and MAP labeling
TLDR
This paper focuses specifically on Nes-terov's optimal first-order optimization scheme for non-smooth convex programs, and shows that in order to obtain an efficiently convergent iteration, this approach should be augmented with a dynamic estimation of a corresponding Lip-schitz constant.
Proximal Alternating Minimization and Projection Methods for Nonconvex Problems: An Approach Based on the Kurdyka-Lojasiewicz Inequality
TLDR
A convergent proximal reweighted l1 algorithm for compressive sensing and an application to rank reduction problems is provided, which depends on the geometrical properties of the function L around its critical points.
Isolated calmness of solution mappings and exact recovery conditions for nuclear norm optimization problems
ABSTRACT We derive some conditions for a feasible point to be the isolated optimal solution for three classes of nuclear norm optimization problems by using the isolated calmness of two set-valued
Hölder-Type Global Error Bounds for Non-degenerate Polynomial Systems
Let F := (f1, …, fp): ℝn → ℝp be a polynomial map, and suppose that S := {x ∈ ℝn : fi(x) ≤ 0,i = 1, …, p}≠∅. Let d := maxi =1, …, p deg fi and ℋ(d,n,p):=d(6d−3)n+p−1.$\mathcal {H}(d, n, p) := d(6d -
Variational analysis of constrained M-estimators
We propose a unified framework for establishing existence of nonparametric M-estimators, computing the corresponding estimates, and proving their strong consistency when the class of functions is
A difference-of-convex approach for split feasibility with applications to matrix factorizations and outlier detection
TLDR
It is observed that, in this setting, the split feasibility problem can be formulated as an optimization problem with a difference-of-convex objective so that standard majorization-minimization type algorithms can be applied.
Nonconvex proximal splitting with computational errors ∗
Throughout this chapter, ‖·‖ denotes the standard Euclidean norm. Problem (1) generalizes the more thoroughly studied class of composite convex optimization problems [30], a class that has witnessed
Kurdyka–Łojasiewicz Exponent via Inf-projection
Kurdyka–Łojasiewicz (KL) exponent plays an important role in estimating the convergence rate of many contemporary first-order methods. In particular, a KL exponent of $$\frac{1}{2}$$ for a suitable
...
1
2
3
4
5
...

References

SHOWING 1-10 OF 759 REFERENCES
Functions of A Single Real Variable
If to each real number x in a certain specified range there corresponds a well-defined number y, then y is said to be a real-valued function of the real variable x over that range, and we write y =
Other results of Poliquin and Rockafellar [1996a], [1996b], are seen in 13.45, 13.46 and 13.49, and in the symmetry assertion of 13.51. The formula in 13.50 follows Poliquin and Rockafellar
  • 1993
Corollary (Lagrange multipliers)
Corollary (Legendre-Fenchel transform as an isometry)
Corollary (Lipschitzian properties under addition)
Corollary (Lipschitzian properties under composition)
Corollary (Pompeiu-Hausdorff distance as a limit)
Corollary (Scorza-Dragoni property)
Corollary (addition of functions)
Corollary (alternative interpretation of normality)
...
1
2
3
4
5
...