Analiza Convexa
Analiza Convexa
The right hand side represents the straight line between (𝑥₁, 𝑓(𝑥₁)) and
(𝑥₂, 𝑓(𝑥₂)) in the graph of 𝑓 as a function of 𝑡; increasing 𝑡 from 0 to 1 or
decreasing 𝑡 from 1 to 0 sweeps this line. Similarly, the argument of the
function 𝑓 in the left hand side represents the straight line between 𝑥₁ and 𝑥₂
in 𝑋 or the 𝑥-axis of the graph of 𝑓. So, this condition requires that the
line between any pair of points on the curve of 𝑓 be above or just meeting the
straight
2. For all 0 < 𝑡 < 1 and all 𝑥₁, 𝑥₂ ∈ 𝑋 such that 𝑥₁ ≠ 𝑥₂:
graph.[2]
The difference of this second condition with respect to the first condition above
(𝑥₁, 𝑓(𝑥₁)) and (𝑥₂, 𝑓(𝑥₂))) between the straight line passing through a
is that this condition does not include the intersection points (for example,
pair of points on the curve of 𝑓 (the straight line is represented by the right
side of this condition) and the curve of 𝑓; the first condition includes the
hand
because 𝑓(𝑥₁) ≤ 𝑓(𝑥₁) and 𝑓(𝑥₂) ≤ 𝑓(𝑥₂) are always true (so not useful to be a
part of a condition).
The second statement can also be modified to get the definition of strict
the strict inequality <. Explicitly, the map 𝑓 is called strictly convex if and
convexity, where the latter is obtained by replacing ≤ with
only if for all real 0 < 𝑡 < 1 and all 𝑥₁, 𝑥₂ ∈ 𝑋 such that
𝑥₁ ≠ 𝑥₂:
A strictly convex function 𝑓 is a function that the straight line between any pair
of points on the curve 𝑓 is above the curve 𝑓 except for
the intersection points between the straight line and the curve.
The function f is said to be concave (resp. strictly concave) if -f(f multiplied by
-1) is convex(resp.strictly convex).
The concept of strong convexity extends and parametrizes the notion of strict
convexity. Intuitively, a strongly-convex function is a
function that grows as fast as a quadratic function.[11] A strongly convex function
- The absolute value function 𝑓(𝑥) = |𝑥| is convex (as reflected in the triangle
derivative is not strictly positive at all points. It is not strongly convex.
-----------------------------------------------------------------------------------
-----------------------------------
**Definition 1.** *The dual space X* of X is the space of linear forms on X with
norm ‖ · ‖*
defined by
**Remark 2.** *For X with norm ‖ · ‖_p, the dual norm is ‖ · ‖_q, where 1/p + 1/q =
1 for p > 1, and
q = ∞ for p = 1. In particular, if X has Euclidean norm, then so does X*.*
**Proof.** (1) ⇒ (2) Let x_λ = x + λ(y − x) for λ ∈ [0, 1]. Using the fundamental
theorem of
calculus and Hölder’s inequality:
and ‖v‖ = 1.
Then
0 ≤ φ(z)
= φ(y) − ⟨∇φ(y), ‖∇φ(y)‖_* / L v⟩ + L/2 ‖ ∇φ(y)‖_* / L v‖²
= f(y) − f(x) − ⟨∇f(x), y − x⟩ − 1/2L ‖∇f(y) − ∇f(x)‖²_*.
(2) ⇒ (5) This follows from the definition of convexity and the inequality in (2).
(5) ⇒ (2) Rewrite (5) as
We prove Jensen’s inequality only for the case where M is a finite set
{m₁, ..., m_k}. Let x_i abbreviates X(m_i) and p_i abbreviates P(m_i). First
consider the case where M contains only two elements. In this case we have
the following.
E_{m~M} [f(X(m))] = p₁f(x₁) + p₂f(x₂)
≥ f(p₁x₁ + p₂x₂)
= f(E_{m~M} [X(m)])
Note that the definition of convexity is simply the statement that Jensen’s
inequality holds for two point distributions. We prove Jensen’s inequality for
finite M by induction on the number of elements of M. Suppose M contains
k elements and assume that Jensen’s inequality holds for distributions on
k − 1 points. We now have the following where the fourth line follows from
the induction hypothesis.
E_{m~P} [f(X(m))]
= p₁f(x₁) + p₂f(x₂) + p₃f(x₃) + ⋯ + p_kf(x_k)
= (p₁ + p₂) ( (p₁ / (p₁ + p₂)) f(x₁) + (p₂ / (p₁ + p₂)) f(x₂) ) + p₃f(x₃) + ⋯
+ p_kf(x_k)
≤ (p₁ + p₂) f ( (p₁ / (p₁ + p₂)) x₁ + (p₂ / (p₁ + p₂)) x₂ ) + p₃f(x₃) + ⋯ +
p_kf(x_k)
≤ f ( (p₁ + p₂) (p₁x₁ + p₂x₂) / (p₁ + p₂) + p₃x₃ + ⋯ + p_kx_k )
= f(p₁x₁ + p₂x₂ + p₃x₃ + ⋯ + p_kx_k)
= f(E_{m~M} [X(m)])