Constrained optimization and Lagrange multipliers
From Wikipedia, the free encyclopedia
| This page is a candidate to be copied to Wikibooks using the Import process. If the page can be re-written into an encyclopedic article, please do so and remove this message. Before you move this content to Wikibooks, verify that it conforms to Wikibooks policies of acceptable content at What is Wikibooks? Often content unacceptable to Wikipedia may not be acceptable on Wikibooks either; facilitate the copying of this article by listing it on Wikibooks:Requests for Import. |
This tutorial presents an introduction to optimization problems that involve finding a maximum or a minimum value of an objective function
subject to a constraint of the form
.
Contents |
[edit] Maximum and minimum
Finding optimum values of the function
without a constraint is a well known problem dealt with in calculus courses. One would normally use the gradient to find stationary points. Then check all stationary and boundary points to find optimum values.
[edit] Example
- f(x,y) = 2x2 + y2
- fx(x,y) = 4x = 0
- fy(x,y) = 2y = 0
f(x,y) has one stationary point at (0,0).
[edit] The Hessian
A common method of determining weather or not a function has an extreme value at a stationary point is to evaluate the hessian[3] of the function at that point. where the hessian is defined as
[edit] Second derivative test
The Second derivative test determines the optimality of stationary point x according to the following rules [2]:
- If H(f) > 0 at point x then f has a local minimum at x
- If H(f) < 0 at point x then f has a local maximum at } x
- If H(f) has negative and positive eigenvalues then x is a saddle point
- Otherwise the test is inconclusive
In the above example.
Therefore f(x,y) has a minimum at (0,0).
[edit] Constrained maximum and minimum
When finding the extreme values of
subject to a constraint
, the stationary points found above will not work. This new problem can be thought of as finding extreme values of
when the point
is restricted to lie on the surface
. The value of
is maximized(minimized) when the surfaces touch each other,i.e , they have a common tangent line. This means that the surfaces gradient vectors at that point are parallel, hence,
The number λ in the equation is known as the Lagrange multiplier.
[edit] Lagrange multiplier method
The Lagrange multiplier methods solves the constrained optimization problem by transforming it into a non-constrained optimization problem of the form:
Then finding the gradient and hessian as was done above will determine any optimum values of
.
Suppose we now want to find optimum values for f(x,y) = 2x2 + y2 subject to x + y = 1 from [2].
Then the Lagrangian method will result in a non-constrained function.
- L(x,y,λ) = 2x2 + y2 + λ(1 − x − y)
The gradient for this new function is
Finding the stationary points of the above equations can be obtained from their matrix from.
This results in x = 1 / 3,y = 2 / 3,λ = 4 / 3.
Next we can use the hessian as before to determine the type of this stationary point.
Since H(L) > 0 then the solution (1 / 3,2 / 3,4 / 3) minimizes f(x,y) = 2x2 + y2 subject to x + y = 1 with f(x,y) = 2 / 3.
[edit] References
- [1] T.K. Moon and W.C. Stirling. Mathematical Methods and Algorithms for Signal Processing. Prentice Hall. 2000.
- [2]http://mat.gsia.cmu.edu/QUANT/NOTES/chap4/node1.html
- [3]http://www.ece.tamu.edu/~chmbrlnd/Courses/ECEN601/ECEN601-Chap3.pdf










