[go: up one dir, main page]

0% found this document useful (0 votes)
4 views17 pages

M01 L08 ConstrainedOptimization

The document provides an overview of constrained optimization, detailing the definition, formulation, and key concepts such as the KKT conditions and primal-dual formulation. It covers the necessary and sufficient conditions for optimization, along with visualizations like contour plots and gradients. Additionally, it highlights various applications of constrained optimization in fields such as machine learning and statistical analysis.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
4 views17 pages

M01 L08 ConstrainedOptimization

The document provides an overview of constrained optimization, detailing the definition, formulation, and key concepts such as the KKT conditions and primal-dual formulation. It covers the necessary and sufficient conditions for optimization, along with visualizations like contour plots and gradients. Additionally, it highlights various applications of constrained optimization in fields such as machine learning and statistical analysis.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 17

Constrained Optimization

Prithwijit Guha
Dept. of EEE, IIT Guwahati
Overview
• The Constrained Optimization Problem
• Contour Plot and Gradients
• Objective Function and Gradients
• The KKT Conditions
• Necessary Conditions
• Sufficient Conditions
• The Primal-Dual Formulation
Constrained Optimization: Definition & Formulation
Constrained Optimization is the Process of Maximizing or Minimizing
a Desired Objective Function while Satisfying a Set of Prevailing
Constraints

𝑀𝑖𝑛𝑖𝑚𝑖𝑧𝑒 𝑓 𝑋

Subject To

𝑔 𝑋 ≤ 0; 𝑖 = 1, … 𝑚 ℎ 𝑋 = 0; 𝑗 = 1, … 𝑙
𝑦
Contour Plot & Gradient

𝑧=6 1
𝛻𝑧 =
1

𝑧 = 𝑓 𝑥, 𝑦 = 𝑥 + 𝑦 − 1
𝑧=2

𝑧=0 𝑧=4 𝑧=8 𝑥


Objective Functions & Gradients: Linear Case
Objective Function
Increases/Decreases
along Gradient
Best Feasible
Solution

Feasible Region
Containing
Regular
Interior Points
Points
Objective Functions & Gradients (Contd.)
𝑀𝑖𝑛𝑚𝑖𝑧𝑒 𝑓(𝑋)
𝛻𝑓(𝑋)
𝑆𝑢𝑏𝑗𝑒𝑐𝑡 𝑡𝑜 𝑔 𝑋 ≤0
𝛻ℎ(𝑋) 𝛻𝑔(𝑋) ℎ 𝑋 =0
Observation:
𝛻𝑓 𝑋 , 𝛻𝑔 𝑋 𝑎𝑛𝑑 𝛻ℎ 𝑋
𝑔(𝑋)
ℎ(𝑋) pass through the same point

𝑓(𝑋) Thus, we can write


𝛻𝑓 𝑋 + 𝜇𝛻𝑔 𝑋 + 𝜆𝛻ℎ 𝑋 = 0
Karush-Kuhn-Tucker (KKT) Conditions
Minimize 𝑓 𝑋 subject to 𝑔 𝑋 ≤ 0; 𝑖 = 1 … 𝑚 and ℎ 𝑋 = 0; 𝑗 = 1 … 𝑙

Construct Lagrangian Function L

𝐿 𝑋; 𝝁, 𝝀 = 𝑓 𝑋 + 𝜇𝑔 𝑋 + 𝜆ℎ 𝑋

Where, 𝝁 = { 𝜇 … 𝜇 … 𝜇 } and 𝝀 = {𝜆 … 𝜆 … 𝜆 }
KKT Conditions (Contd.)

𝐎𝐩𝐭𝐢𝐦𝐚𝐥𝐢𝐭𝐲: 𝛻 𝐿 𝑋; 𝝁, 𝝀 = 𝛻 𝑓 𝑋 + 𝜇𝛻 𝑔 𝑋 + 𝜆𝛻 ℎ 𝑋 =0

𝐍𝐨𝐧 − 𝐧𝐞𝐠𝐚𝐭𝐢𝐯𝐢𝐭𝐲: 𝜇 ≥ 0 ; 𝑖 = 1 … 𝑚

𝐂𝐨𝐦𝐩𝐥𝐞𝐦𝐞𝐧𝐭𝐚𝐫𝐢𝐭𝐲: 𝜇 𝑔 𝑋 = 0; 𝑖 = 1, … 𝑚

𝐅𝐞𝐚𝐬𝐢𝐛𝐢𝐥𝐢𝐭𝐲: 𝑔 𝑋 ≤ 0; 𝑖 = 1, … 𝑚 and ℎ 𝑋 = 0; 𝑗 = 1, … 𝑙

Necessary Conditions
KKT Conditions (Contd.)
𝑋 ∗ : Optimal Solution of 𝛻 𝐿 𝑋; 𝝁, 𝝀 = 0

Hessian Matrix 𝛻 𝐿 𝑋 ∗ ; 𝜇, 𝜆 𝑖𝑠 𝑃𝑜𝑠𝑖𝑡𝑖𝑣𝑒 𝐷𝑒𝑓𝑖𝑛𝑖𝑡𝑒 𝑖𝑛 𝑻 ⊂ ℝ

𝑻 = 𝑦: 𝛻𝑔 𝑋 ∗ 𝑦 = 0; 𝑖 = 1 … 𝑚 ∧ 𝛻ℎ 𝑋 ∗ y = 0; j = 1 … 𝑙

𝐒𝐮𝐟𝐟𝐢𝐜𝐢𝐞𝐧𝐭 𝐂𝐨𝐧𝐝𝐢𝐭𝐢𝐨𝐧: 𝒚𝑻 𝛻 𝐿 𝑋 ∗ ; 𝜇, 𝜆 y > 0; 𝑦 ≠ 0; 𝑦 ∈ 𝑻

𝐒𝐭𝐫𝐨𝐧𝐠𝐞𝐫 𝐂𝐨𝐧𝐝𝐢𝐭𝐢𝐨𝐧: 𝒚𝑻 𝛻 𝐿 𝑋 ∗ ; 𝜇, 𝜆 y > 0; 𝑦 ≠ 0; 𝑦 ∈ ℝ𝒏


Problem

Minimize 𝑓 𝑥 , 𝑥 = 2𝑥 + 𝑥

Subject To

ℎ 𝑥 ,𝑥 = 𝑥 +𝑥 −1 =0
Solution
𝐿 𝑥 , 𝑥 ; 𝜆 = 2𝑥 + 𝑥 + 𝜆(𝑥 + 𝑥 − 1)

𝜕𝐿 1 𝜕𝐿 1
=0→𝑥 =− =0→𝑥 =−
𝜕𝑥 𝜆 𝜕𝑥 2𝜆
√5
𝜆= ±
2

ℎ(𝑥 , 𝑥 ) B A= 𝑥 =−
2
,𝑥 = −
1
5 5

𝑓(𝑥 , 𝑥 ) 2 1
A B= 𝑥 =
5
,𝑥 =
5
𝑓(𝑥 , 𝑥 )
Primal-Dual Formulation
Minimize 𝑓 𝑋 subject to 𝑔 𝑋 ≤ 0; 𝑖 = 1 … 𝑚 and ℎ 𝑋 = 0; 𝑗 = 1 … 𝑙

𝑋 ∗ , 𝝁∗ , 𝝀∗ : Optimal Solution of the 𝐏𝐫𝐢𝐦𝐚𝐥 𝐏𝐫𝐨𝐛𝐥𝐞𝐦 𝛻 𝐿 𝑋; 𝝁, 𝝀 = 0

(𝑋 ∗ , 𝝁∗ , 𝝀∗ ) is a 𝐒𝐚𝐝𝐝𝐥𝐞 𝐏𝐨𝐢𝐧𝐭 of 𝐿 𝑋; 𝝁, 𝝀 i.e.


𝐿(𝑋 ∗ , 𝝁, 𝝀) ≤ 𝐿(𝑋 ∗ , 𝝁∗ , 𝝀∗ ) ≤ 𝐿(𝑋, 𝝁∗ , 𝝀∗ )

(𝑋 ∗ , 𝝁∗ , 𝝀∗ ) 𝐬𝐨𝐥𝐯𝐞𝐬 𝐭𝐡𝐞 𝐟𝐨𝐥𝐥𝐨𝐰𝐢𝐧𝐠 𝐝𝐮𝐚𝐥 𝐩𝐫𝐨𝐛𝐥𝐞𝐦

𝑀𝑎𝑥𝑖𝑚𝑖𝑧𝑒 𝐿 = 𝑓 + 𝝁 𝒈 + 𝝀 𝒉
𝑆𝑢𝑏𝑗𝑒𝑐𝑡 𝑇𝑜 𝛁𝐿 = 0 ∧ [𝝁 ≥ 0]
Problem
Minimize 𝑓 𝑥 , 𝑥 = 𝑥 −3 + 𝑥 −3

𝐒𝐮𝐛𝐣𝐞𝐜𝐭 𝐓𝐨

g 𝑥 ,𝑥 = 2𝑥 + 𝑥 − 2 ≤ 0

g 𝑥 ,𝑥 = −𝑥 ≤ 0
g 𝑥 ,𝑥 = −𝑥 ≤ 0
Solution
𝐿 𝑥 ,𝑥 ;𝝁 = 𝑥 − 3 + 𝑥 −3 + 𝜇 2𝑥 + 𝑥 − 2 − 𝜇 𝑥 − 𝜇 𝑥

𝜕𝐿 𝜇 𝜕𝐿 𝜇 𝜇
=0→𝑥 = −𝜇 +3 =0→𝑥 = − +3
𝜕𝑥 2 𝜕𝑥 2 2

5𝜇 𝜇 𝜇 𝜇 𝜇
𝑄 𝝁 =− − − + + 𝜇 𝜇 + 7𝜇 − 3𝜇 − 3𝜇
4 4 4 2
Dual Function to be Maximized w. r. t. 𝝁
Applications
• Principal Component Analysis
• Linear Discriminant Analysis
• Factor Analysis
• Regularization
• Neural Network Learning
• Support Vector Machines
• Parameter Estimation And Many Others…
Summary
• Constrained Minimization
• Visualizations
• Contour Plots & Gradients
• Objective Functions & Gradients
• Feasible Regions and Best Solutions
• Karush-Kuhn-Tucker Conditions
• Necessary Conditions
• Sufficient Conditions
• Primal-Dual Formulations
• Applications

You might also like