0% found this document useful (0 votes)
49 views26 pages

Chapter 5.3

The document discusses necessary conditions for constrained optimization problems. It introduces concepts such as active constraints, regular points, tangent cones, and the Karush-Kuhn-Tucker (KKT) conditions. The key points are: 1) The KKT conditions provide first-order necessary conditions for a point to be a local minimum of a constrained optimization problem. They involve the gradient of the objective function and gradients of active constraints. 2) For a regular point (where active constraint gradients are independent), the KKT conditions are also sufficient for a minimum within the set of active constraints. 3) Second-order sufficient conditions involve the augmented Hessian and concepts like the tangent cone and a subset called the tangent

Uploaded by

data science
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
49 views26 pages

Chapter 5.3

The document discusses necessary conditions for constrained optimization problems. It introduces concepts such as active constraints, regular points, tangent cones, and the Karush-Kuhn-Tucker (KKT) conditions. The key points are: 1) The KKT conditions provide first-order necessary conditions for a point to be a local minimum of a constrained optimization problem. They involve the gradient of the objective function and gradients of active constraints. 2) For a regular point (where active constraint gradients are independent), the KKT conditions are also sufficient for a minimum within the set of active constraints. 3) Second-order sufficient conditions involve the augmented Hessian and concepts like the tangent cone and a subset called the tangent

Uploaded by

data science
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 26

E1

 251  Linear  and  Nonlinear  


Op2miza2on  
 
   Chapter  5:  Condi/ons  of  constrained    
maxima  and  minima      

5.1:    Geometric  characteriza/on  


5.2:    Algebraic  characteriza/on  for  equality    
                 constraints  
5.3:    Algebraic  characteriza/on  for  inequality    
                 constraints  
 
 
 
5.3  Problem  with  inequality  constraints  
Minimize f (x)
2 Chapter
Subject 11 Constrained
to h(x) Minimization
= 0 and g(x) ≤0 Conditions

h(x) : m × 1 vector function


irst-Order × 1 vector function
g(x) : p Necessary Conditions
Definition: Active constraint
ith the following generalization of our previous definition it is possible to parallel
A constraint
e development gi (x) ≤ 0conditions
of necessary is called active *
gi (x ) = 0.
*
at x if constraints.
for equality
Definition. Let x∗ be a point satisfying the constraints

h!x∗ " = 0# g!x∗ " ! 0# (E1)


(33)

and let J be the set of indices j for which gj !x∗ " = 0. Then x∗ is said to be a
(E1) if the gradient vectors !hi !x∗ ", !gj !x∗ ",
regular point of the constraints (33)
1 ! i ! m# j ∈ J are linearly independent.
We note that, following the definition of active constraints given in
ction 11.1, a point x∗ is a regular point if the gradients of the active constraints

Some notations:
Ω : set of points satisfying h(x) = 0 and g(x) ≤ 0
Ω E : set of points satisfying h(x) = 0, g(x) ≤ 0 and
g i (x) = 0 for all i in J.
Ω N : set of points satisfying h(x) = 0, g(x) ≤ 0 and
gi (x) < 0 for at least one i ∈J.
Ω = ΩE + ΩN .

If x* is a local minimum of f (x) within Ω, it should be also a local


minimum within Ω E .
For a candidate point x* , we at looking at the following types
of curves:

(1) SE is the set of curves that are entirely in Ω E , which pass through
the point x* . Each curve can be split into two curves
that end at x* through opposing directions. Let SE be the set of such
curves.
(2) SN , set of curves in Ω but not in Ω E (except x* ), which end at the
point x* .
(3) S is the set of all curves in Ω that end at x* .
S = SE + S N .
Tangents of curves in S is the tangent cone TΩ (x* ).

Tangents of curves in SE is the cone T(x* ), and T(x* ) ⊂ TΩ (x* ).

Since each curve in SE has an opposing curve, for each d ∈ T(x* ),


-d ∈T(x* ).

A cone satifying this property is a subspace.


5.3.1  
tion 11.1, FONC  
a point for  
x∗ inequality   constraints  
is a regular point if the gradients of the active constraints
linearly independent. Or, equivalently, x∗ is regular for the constraints if it is
ular in the sense of the earlier definition for equality constraints applied to the
veTheorem  
constraints. 5.3A  

Karush–Kuhn–Tucker Conditions. Let x∗ be a relative minimum point for the


problem

minimize f!x"
(34)
(E2)
subject to h!x" = 0# g!x" ! 0#

and suppose x∗ is a regular point for the constraints. Then there is a vector
"∈E pp with # " 0 such that
mm and a vector # ∈ E

∇f (x *
!f!x ) + ∗J"h+ )λ!h!x
* T
(x" ∗ * )µ =0
+ J g (x" + #T !g!x∗ " = 0 (35)
(E3)

#T g!x∗ " = 0$ (36)


(E4)
Proof:.
•(E4) imply that µi = 0 when gi is not active.

•(E3) denotes FONC for a minimum within Ω E .


Hence it is also FONC for minimum in Ω.

• Consider a curve x(t) ∈SN such that x(0) = x* . Let x(t) is such
that bi (t) = gi (x(t)) satisfies bi (t) < 0 when t > 0 for exactly one
i ∈J and, for remaining j ∈J, b j (t) = g j (x(t)) satisfies b j (t) = 0.
for all t.

•The above statement implies that bi′(0) ≤ 0 for one i ∈J and b ′j (0) = 0
remaining j ∈J. Hence ( x! (0)) ∇gi (x* ) ≤ 0 for one i ∈J and
T

( x! (0))T ∇g j (x* ) = 0 for remaining j ∈J.


• Note that ( x! (0)) ∇hk (x* ) = 0 for all ks.
T

• Let d = x! (0). Premultiplying (E3) with dT gives


dT ∇f (x* ) + µi dT ∇gi (x* ) = 0.

• Since we know that dT ∇gi (x* ) is non-positive and dT ∇f (x* )


can not be negative, the above equation imply that µi is non-negative.
5.3.2  Varia/ons  

Maximize f (x)
subject to h(x) = 0 and g(x) ≤ 0 KKT conditions:
KKT conditions: ∇f (x) + J h (x* )λ * + J g (x* )µ * = 0
−∇f (x) + J h (x* )λ * + J g (x* )µ * = 0 µ* ≤ 0
µ* ≥ 0 µ *T g(x* ) = 0
µ *T g(x* ) = 0 h(x* ) = 0
h(x* ) = 0 g(x* ) ≤ 0
g(x* ) ≤ 0
Minimize f (x)
subject to h(x) = 0 and g(x) ≥ 0 KKT conditions:
KKT conditions: ∇f (x) + J h (x* )λ * + J g (x* )µ * = 0
∇f (x) + J h (x* )λ * − J g (x* )µ * = 0 µ* ≤ 0
µ* ≥ 0 µ *T g(x* ) = 0
µ *T g(x* ) = 0 h(x* ) = 0
h(x* ) = 0 g(x* ) ≥ 0
g(x* ) ≥ 0
Maximize f (x)
subject to h(x) = 0 and g(x) ≥ 0 KKT conditions:
KKT conditions: ∇f (x) + J h (x* )λ * + J g (x* )µ * = 0
−∇f (x) + J h (x* )λ * − J g (x* )µ * = 0 µ* ≥ 0
µ* ≥ 0 µ *T g(x* ) = 0
µ *T g(x* ) = 0 h(x* ) = 0
h(x* ) = 0 g(x* ) ≥ 0
g(x* ) ≥ 0
Example:
Maximize/minimize f (x)
subject to g(x) ≥ 0
KKT conditions for x1 (maximum):
∇f (x 1 ) + ∇g(x 1 )µ1 = 0
µ1 ≥ 0
µ1T g(x 1 ) = 0
g(x 1 ) ≥ 0
KKT conditions for x 2 (minimum):
∇f (x 2 ) + ∇g(x 2 )µ2 = 0
µ2 ≤ 0
µ2T g(x 2 ) = 0
g(x 2 ) ≥ 0
5.3.3  Minimiza/on   with  posi/vity  
The above example is a specialconstraints  
case of a more general prob

The KKT condition for this problem has the form

From the above, we can eliminate // to obtain


5.3.4  Second  order  condi/ons  
Recap − tangent cone:
A vector d said to be a tangent to the constraint set Ω at point x* ,
if there is a feasible sequence {x k } approaching x* , and sequence
x k − x*
of positive scalars {t k } approaching zero such that lim = d.
k→∞ tk
The set of all tangents to Ω at point x* is called the tangent cone
TΩ (x* ).

Definition M Ω (x* ):
For a feasible point x* , the set is defined as
⎧ T
∇h *
) = 0, i = 1,...., m
⎪ d i (x
M Ω (x ) = ⎨d T
*

⎪⎩ d ∇g j (x *
) ≤ 0, for all j s.t. g j (x *
)=0

Theorem 5.3B : For a regular point x* ,M Ω (x* ) = TΩ (x* ).


m p
The augmented Hessian: L(x, λ , µ ) = F(x) + ∑ λi Hi (x) + ∑ µi Gi (x)
i=1 i=1

The set J p (x* , µ * ) :


The set of indices such that for every j ∈ J p (x* , µ * ), g j (x* ) = 0
and µ * j > 0.

The set J z (x* , µ * ) :


The set of indices such that for every j ∈ J z (x* , µ * ), g j (x* ) = 0
and µ * j = 0.

J(x* , µ * ) = J p (x* , µ * ) + J z (x* , µ * ).


Spliting the set M Ω (x* ) :
M Ω (x* ) = M(x* ) + M1 (x* ) + M 2 (x* )

⎧ y ∈! n : yT ∇hi (x* ) = 0, ∀i = 1,..., m; ⎫


⎪ ⎪
M(x* ) = ⎨ :yT ∇gi (x* ) = 0,∀i ∈J p (x* , µ * ) ⎬
⎪ * ⎪
⎩ :y T
∇gi (x *
) = 0,∀i ∈J z (x *
, µ )⎭

For a regular point x* , M(x* ) = T(x* ), the tangent subspace.


The subset M1 (x* ) :
⎧ y ∈! n : yT ∇hi (x* ) = 0,∀i = 1,..., m; ⎫
⎪ ⎪
⎪ :y ∇gi (x ) ≤ 0,∀i ∈J(x , µ ),
T * * *

M1 (x ) = ⎨
*
* ⎬
⎪ : y ∇gi (x ) < 0, for at least one i ∈J p (x , µ ) ⎪
T * *

⎪ :y T
∇g (x *
) = 0,∀i ∈J (x *
, µ *
) ⎪
⎩ i z ⎭

For a regular point x* , M1 (x* ) is the set of tangents of the curves x(t)
satisfying x(0) = x* and hi (x(t)) = 0,∀i = 1,..., m,
gi (x(t)) ≤ 0,∀i ∈J(x* , µ * ),
gi (x(t)) < 0, for at least one i ∈J p (x* , µ * ) for t > 0.
gi (x(t)) = 0,∀i ∈J z (x* , µ * )
The subset M 2 (x* ) :
⎧ y ∈! n : yT ∇hi (x* ) = 0, ∀i = 1,..., m; ⎫
⎪ ⎪
⎪ :y T
∇gi (x *
) ≤ 0, ∀i ∈J(x* , µ * ) ⎪
M 2 (x ) = ⎨
*
* ⎬
⎪ :y T
∇gi (x *
) < 0, for at least one i ∈J z (x , µ ) ⎪
*

⎪ :y T
∇g (x *
) = 0, ∀i ∈J p (x* , µ * ) ⎪
⎩ i ⎭

For a regular point x* , M 2 (x* ) is the set of tangents of the curves x(t)
satisfying x(0) = x* and hi (x(t)) = 0,∀i = 1,..., m,
gi (x(t)) ≤ 0,∀i ∈J(x* , µ * ),
gi (x(t)) < 0, for at least one i ∈J z (x* , µ * ) for t > 0.
gi (x(t)) = 0,∀i ∈J p (x* , µ * )
Second order necessary condition (Theorem 5.3B):
Let x* be the minimum of the function f (x) in ! n subject to
constraint h(x) = 0 and g(x) ≤ 0 where h(x) and g(x) are m × 1
and p × 1 vector functions in ! n . Also, let x* be the regular
point of the constraints. Then there exists vectors λ * ∈! m ,
µ * ∈! p , and x* ∈! n satisfying KKT conditions. Also,
for every vector y in the subset M(x* ) + M 2 (x* ), we have
yT L(x* , λ * , µ * )y ≥ 0.
Second order sufficient condition (Theorem 5.3C):
For a function f (x), and constraints h(x) = 0 and g(x) ≤ 0,
let x* ∈! n , λ * ∈! m and µ * ∈! p be the vectors satisfying
KKT conditions. For every vector y in the subset
M(x* ) + M 2 (x* ), if we have yT L(x* , λ * , µ * )y > 0, then x* is a
strict minimum of the function f (x) subject to constraints h(x) = 0
and g(x) ≤ 0.
Second  order  condi/ons  are  proved  using  the  following  
 theorem  

Theorem 5.3D:
Consider the function f (x) in ! n and constraints h(x) = 0 and
g(x) ≤ 0 where h(x) and g(x) are m × 1 and p × 1 vector functions
in ! n . Let there exist vectors, λ * ∈! m , µ * ∈! p , and
x* ∈! n such that ∇f (x* ) + J h (x* )λ * + J g (x* )µ * = 0, µ * ≥ 0,
and µ *T g(x* ) = 0. Then for every s(t) = f (x(t)) such that
x'(0) = d ∈M (x* ) + M 2 (x* ), and x(0) = x* we have
s(ε) = s(0) + ε 2 dT L(x* , λ * , µ * )d + o(ε 2 ).
Proof of Theorem 5.3D:
For the feasible curve, define
m p
c(t) = L(x(t), λ * , µ * ) = f (x(t)) + ∑ λ *i hi (x(t)) + ∑ µ *i gi (x(t))
i=1 i=1

Differentiating gives
⎛ m p

c′(t) = x ′ (t) ⎜ ∇f (x(t)) + ∑ λ i ∇hi (x(t)) + ∑ µ i ∇gi (x(t))⎟
T * *

⎝ ⎠
!########"########$
i=1 i=1

L x (x(t ),λ * , µ * )

⎛ m p

c′′(t) = x ′′ (t) ⎜ ∇f (x(t)) + ∑ λ i ∇hi (x(t)) + ∑ µ i ∇gi (x(t))⎟ +
T * *

⎝ ⎠
!########"########$
i=1 i=1

L x (x(t ),λ * , µ * )

⎛ m p

x ′ (t) ⎜ F(x(t)) + ∑ λ i Hi (x(t)) + ∑ µ i Hi (x(t))⎟ x ′(t)
T * *

⎝ ⎠
!####### i=1
#"######## i=1
$
L xx (x(t ),λ * , µ * )
Now by definition
c(0) = f (x(0)) =L(x* , λ * , µ * ) = f (x* )

By KKT, L x (x(0), λ * , µ * ) =L x (x* , λ * , µ * ) = 0.

Hence
c′(0) =0 c′′(0) =x ′T (0)L xx (x* , λ * , µ * )x ′(0)

Substituting in second order Taylor's formula


c(ε) = f (x* ) +0.5 x ′T (0)L xx (x(0), λ * , µ * )x ′(0) + o(ε 2 ).
Since x(t) was defined such that gi (x(t)) < 0 for only those
i ' s in J for which the corresponding µi is zeros, c(ε) = s(ε),

Recap:
m p
c(ε) = L(x(ε), λ * , µ * ) = f (x(ε)) + ∑ λ *i hi (x(ε)) + ∑ µ *i gi (x(ε))
i=1 i=1

s(ε) = f (x(ε)).

Hence
s(ε) = f (x* ) +0.5 x ′T (0)L xx (x(0), λ * , µ * )x ′(0) + o(ε 2 )

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy