Segment parametrization difference from the left endpoint.
Proposition 1.3.1.
Let Q be a closed convex subset of a finite-dimensional real normed space and let f : Q → Real
be convex and differentiable. If the gradient is Lipschitz with constant L > 0 in the dual norm
(equation (grad_lipschitz)), then for any x, y in Q,
f y <= f x + pairing (grad f x) (y - x) + (L / 2) (norm (y - x))^2
(inequality (3.1)).
Definition 1.3.1.
Let f satisfy the assumptions of Proposition 1.3.1. For x ∈ Q, define T_Q(x) ∈ Q to be any
minimizer of y ↦ ⟪∇ f(x), y - x⟫ + (L / 2) ‖y - x‖^2 over y ∈ Q
(equation (3.2)). If the norm is not strictly convex, the minimizer may be non-unique, and
T_Q(x) denotes an arbitrary choice.
Equations
Instances For
Existence of a minimizer for the quadratic model on a closed set.
If a minimizer exists, T_Q is a minimizer of the quadratic model.
The quadratic model at T_Q equals the infimum over Q.
Proposition 1.3.2.
Under the assumptions of Proposition 1.3.1, for any x ∈ Q we have
f(T_Q(x)) ≤ f(x) + min_{y ∈ Q} { ⟪∇ f(x), y - x⟫ + (L / 2) ‖y - x‖^2 }
(equation (3.3)).
Definition 1.3.2.
Let Q ⊆ E be closed and convex. A function d : E → ℝ is a prox-function for Q if it is
continuous and σ-strongly convex on Q for some σ > 0.
Equations
- ProxFunction Q d = IsProxFunction Q d
Instances For
Definition 1.3.2.
Define the prox-center x0 ∈ Q by x0 ∈ argmin_{x ∈ Q} d x (equation (prox_center_3)) and
assume the normalization d x0 = 0.
Equations
- IsNormalizedProxCenter Q d x0 = (IsProxCenter Q d x0 ∧ d x0 = 0)
Instances For
Proposition 1.3.3.
Assume d is σ-strongly convex on Q and x0 ∈ argmin_{x ∈ Q} d x with d x0 = 0. Then for
all x ∈ Q, d x ≥ (σ / 2) ‖x - x0‖^2 (equation (3.4)).
Definition 1.3.3.
Consider the smooth convex minimization problem min_{x ∈ Q} f x (equation (3.5)), where
Q ⊆ E is closed and convex, and f is convex and differentiable on Q with L-Lipschitz
continuous gradient as in (grad_lipschitz).
Equations
- One or more equations did not get rendered due to their size.
Instances For
Definition 1.3.4.
Given points x_0, …, x_k ∈ Q, define
ψ_k = min_{x ∈ Q} { (L/σ) d x + ∑_{i=0}^k α_i [ f(x_i) + ⟪∇ f(x_i), x - x_i⟫ ] }
(equation (psi_k_def)).
Equations
- One or more equations did not get rendered due to their size.
Instances For
Definition 1.3.4.
Given a sequence {y_k} ⊆ Q, define the relation
(R_k): A_k f(y_k) ≤ ψ_k (equation (Rk)).
Instances For
The quadratic model at T_Q is minimal on Q.
Scale the prox-center lower bound to compare with the quadratic model.
Pointwise lower bound for the psi_0 comparison.
Proposition 1.3.4.
Assume α_0 ∈ (0,1]. Let x0 be the prox-center of Q from Definition 3.4, and set
y0 = T_Q(x0) (equation (y0_def)). Then the relation (R_0) in (Rk) holds.
Definition 1.3.5.
For k ≥ 0, define z_k ∈ Q to be any minimizer achieving ψ_k in (psi_k_def), namely
z_k ∈ argmin_{x ∈ Q} { (L/σ) d x + ∑_{i=0}^k α_i [ f(x_i) + ⟪∇ f(x_i), x - x_i⟫ ] }
(equation (zk_def)).
Equations
- One or more equations did not get rendered due to their size.
Instances For
Supporting hyperplane inequality for a differentiable convex function on a convex set.
Change-of-variables identity for y = τ_k x + (1-τ_k) y_k.
Combine the gradient pairing terms using the update for x_{k+1}.
Quadratic-model identity after a change of variables.
The quadratic model at T_Q dominates the function decrease.
Affine function x ↦ ⟪s, x - x0⟫ is convex on a convex set.