L10_Subgrad_PGD (partially annotated)
L10_Subgrad_PGD (partially annotated)
L10_Subgrad_PGD (partially annotated)
x ∈ X ⊆ Rn
Assumptions:
X is a compact set in Rn
f : Rn → R, g : Rn → Rm , and h : Rn → Rp are continuous
Problem setting
x ∈ X ⊆ Rn
Assumptions:
X is a compact set in Rn
f : Rn → R, g : Rn → Rm , and h : Rn → Rp are continuous
For solving dual, consider gradient ascent on θ:
[λ; µ](k+1) = [λ; µ](k) + tk ∇θ λ(k) , µ(k)
tk : step length
·) is often not differentiable!
Difficulty: θ(·,
5λ − 4 if λ ≤ −1
Eg. θ(λ) = λ − 8 if − 1 ≤ λ ≤ 2 in example 10.11.
−3λ if λ ≥ 2
Another motivating example
Definition 10.1
Suppose
S ⊆ Rn be a nonempty convex set
f : S → R be a convex function
A vector ξ ∈ Rn is a subgradient of f at x̄ ∈ S if
Definition 10.1
Suppose
S ⊆ Rn be a nonempty convex set
f : S → R be a convex function
A vector ξ ∈ Rn is a subgradient of f at x̄ ∈ S if
Proposition 1
∂f is a convex set. If f is differentiable at x, then
Proposition 1
∂f is a convex set. If f is differentiable at x, then
Proposition 2
If f is continuous and convex, then minx∈Rn f (x) is attained at x∗ if and
only if 0 ∈ ∂f (x∗ ).
Some useful propositions
Proposition 1
∂f is a convex set. If f is differentiable at x, then
Proposition 2
If f is continuous and convex, then minx∈Rn f (x) is attained at x∗ if and
only if 0 ∈ ∂f (x∗ ).
Proposition 3
Let f , g be two convex functions. Under some mild conditions (which
assumed to be true in this course), the subdifferential of f + g is given by
Definition 10.3
Given a set S, C = conv (S) is the convex hull of S, if C is the smallest
convex set that contains S.
Convex hull
Definition 10.3
Given a set S, C = conv (S) is the convex hull of S, if C is the smallest
convex set that contains S.
Proposition 4
If S = {v1 , · · · , vn }, then
n n
X n
X o
conv (S) = v = λi vi , λi ≥ 0, λi = 1
i=1 i=1
Proposition 5
Suppose f (x) = max{f1 (x), · · · , fm (x)}, where fi are all convex and
continuously differentiable functions. Suppose
f (x∗ ) = f1 (x∗ ) = · · · = fj (x∗ ). Then
Running through all cases, the only possibility where 0 ∈ ∂θ(λ∗ ) is when λ∗ = 2.
Example 10.6
Find the subgradient of
The projected gradient method is a method that pulls the iterate back to
the feasible region, using the basic template of the usual gradient descent
algorithm.
Projection on a convex set
Theorem 10.7 (Projection theorem)
Let C be a closed convex set in Rn .
(a) For every z ∈ Rn , there exists a unique minimizer (denoted as ΠC (z)
and called as the projection of z onto C ) of
n1 o
min kx − zk2 | x ∈ C
2
where k · k is the Euclidean norm.
Projection on a convex set
Theorem 10.7 (Projection theorem)
Let C be a closed convex set in Rn .
(a) For every z ∈ Rn , there exists a unique minimizer (denoted as ΠC (z)
and called as the projection of z onto C ) of
n1 o
min kx − zk2 | x ∈ C
2
where k · k is the Euclidean norm.
(b) x∗ := ΠC (z) is the projection of z onto C if and only if
hz − x∗ , x − x∗ i ≤ 0 ∀ x ∈ C .
Projection on a convex set
Theorem 10.7 (Projection theorem)
Let C be a closed convex set in Rn .
(a) For every z ∈ Rn , there exists a unique minimizer (denoted as ΠC (z)
and called as the projection of z onto C ) of
n1 o
min kx − zk2 | x ∈ C
2
where k · k is the Euclidean norm.
(b) x∗ := ΠC (z) is the projection of z onto C if and only if
hz − x∗ , x − x∗ i ≤ 0 ∀ x ∈ C .
[Step k] For k = 0, 1, 2, 3 · · · ,
(a) d(k) := −∇f (x(k) ).
(b) if kd(k) k < , stop.
(c) else,
(i) Choose fixed tk , or solve
[Step 0] Select x(0) , and > 0. [Step 0] Select x(0) , and > 0.