Documentation

Papers.SmoothMinimization_Nesterov_2004.Sections.section02_part3

axiom smoothedMaxFunction_hasFDerivAt {E1 : Type u_1} {E2 : Type u_2} [NormedAddCommGroup E1] [NormedSpace E1] [FiniteDimensional E1] [NormedAddCommGroup E2] [NormedSpace E2] [FiniteDimensional E2] (Q2 : Set E2) (A : E1 →L[] E2 →L[] ) (phihat d2 : E2) (μ : ) ( : E1E2) :
let := SmoothedMaxFunction Q2 A phihat d2 μ; let A' := { toFun := fun (x : E1) => (A x), map_add' := , map_smul' := }; ∀ (x : E1), HasFDerivAt (LinearMap.toContinuousLinearMap ((AdjointOperator A') ( x))) x

Assumed differentiability of the smoothed max-function (Danskin-type formula).

axiom smoothedMaxFunction_gradient_lipschitz {E1 : Type u_1} {E2 : Type u_2} [NormedAddCommGroup E1] [NormedSpace E1] [FiniteDimensional E1] [NormedAddCommGroup E2] [NormedSpace E2] [FiniteDimensional E2] (Q2 : Set E2) (A : E1 →L[] E2 →L[] ) (phihat d2 : E2) (μ σ2 : ) ( : E1E2) :
let A' := { toFun := fun (x : E1) => (A x), map_add' := , map_smul' := }; ∃ ( : ), = 1 / (μ * σ2) * OperatorNormDef A' ^ 2 LipschitzWith .toNNReal fun (x : E1) => LinearMap.toContinuousLinearMap ((AdjointOperator A') ( x))

Assumed Lipschitz continuity of the smoothed max-function gradient.

axiom smoothedMaxFunction_contDiff {E1 : Type u_1} {E2 : Type u_2} [NormedAddCommGroup E1] [NormedSpace E1] [FiniteDimensional E1] [NormedAddCommGroup E2] [NormedSpace E2] [FiniteDimensional E2] (Q2 : Set E2) (A : E1 →L[] E2 →L[] ) (phihat d2 : E2) (μ : ) :
ContDiff (↑1) (SmoothedMaxFunction Q2 A phihat d2 μ)

Assumed C^1 smoothness of the smoothed max-function.

theorem smoothedMaxFunction_key_inequality {E1 : Type u_1} {E2 : Type u_2} [NormedAddCommGroup E1] [NormedSpace E1] [FiniteDimensional E1] [NormedAddCommGroup E2] [NormedSpace E2] [FiniteDimensional E2] (Q2 : Set E2) (A : E1 →L[] E2 →L[] ) (phihat d2 : E2) (μ σ2 : ) ( : 0 < μ) (hconv : StrongConvexOn Q2 σ2 d2) (hphi : ConvexOn Q2 phihat) ( : E1E2) (hmax : ∀ (x : E1), IsSmoothedMaximizer Q2 A phihat d2 μ x ( x)) :
have A' := { toFun := fun (x : E1) => (A x), map_add' := , map_smul' := }; ∀ (x1 x2 : E1), μ * σ2 * x1 - x2 ^ 2 DualPairing ((AdjointOperator A') ( x1 - x2)) (x1 - x2)

Key inequality for the smoothed maximizers, assuming convexity of phihat.

theorem smoothedMaxFunction_properties {E1 : Type u_1} {E2 : Type u_2} [NormedAddCommGroup E1] [NormedSpace E1] [FiniteDimensional E1] [NormedAddCommGroup E2] [NormedSpace E2] [FiniteDimensional E2] (Q2 : Set E2) (A : E1 →L[] E2 →L[] ) (phihat d2 : E2) (μ σ2 : ) ( : 0 < μ) (hσ2 : 0 < σ2) (hconv : StrongConvexOn Q2 σ2 d2) ( : E1E2) (hmax : ∀ (x : E1), IsSmoothedMaximizer Q2 A phihat d2 μ x ( x)) :
have := SmoothedMaxFunction Q2 A phihat d2 μ; have A' := { toFun := fun (x : E1) => (A x), map_add' := , map_smul' := }; ContDiff (↑1) ConvexOn Set.univ (∀ (x : E1), HasFDerivAt (LinearMap.toContinuousLinearMap ((AdjointOperator A') ( x))) x) ∃ ( : ), = 1 / (μ * σ2) * OperatorNormDef A' ^ 2 LipschitzWith .toNNReal fun (x : E1) => LinearMap.toContinuousLinearMap ((AdjointOperator A') ( x))

Theorem 1.2.1. Assume μ > 0 and that d2 is σ2-strongly convex on Q2. Then f_μ defined by (2.5) is well-defined and continuously differentiable for all x ∈ E1. Moreover, f_μ is convex and ∇ f_μ(x) = A* u_μ(x) (equation (2.6)). The gradient is Lipschitz continuous with constant L_μ = (1/(μ σ2)) ‖A‖_{1,2}^2 (equation (eq:L_mu)).