Insights Blog
-- Browse All Articles --
Physics Articles
Physics Tutorials
Physics Guides
Physics FAQ
Math Articles
Math Tutorials
Math Guides
Math FAQ
Education Articles
Education Guides
Bio/Chem Articles
Technology Guides
Computer Science Tutorials
Forums
Trending
Featured Threads
Log in
Register
What's new
Search
Search
Google search
: add "Physics Forums" to query
Search titles only
By:
Latest activity
Register
Menu
Log in
Register
Navigation
More options
Contact us
Close Menu
JavaScript is disabled. For a better experience, please enable JavaScript in your browser before proceeding.
You are using an out of date browser. It may not display this or other websites correctly.
You should upgrade or use an
alternative browser
.
Forums
Lagrange multiplier
Recent contents
View information
Top users
Description
In mathematical optimization, the method of Lagrange multipliers is a strategy for finding the local maxima and minima of a function subject to equality constraints (i.e., subject to the condition that one or more equations have to be satisfied exactly by the chosen values of the variables). It is named after the mathematician Joseph-Louis Lagrange. The basic idea is to convert a constrained problem into a form such that the derivative test of an unconstrained problem can still be applied. The relationship between the gradient of the function and gradients of the constraints rather naturally leads to a reformulation of the original problem, known as the Lagrangian function.The method can be summarized as follows: in order to find the maximum or minimum of a function
f
(
x
)
{\displaystyle f(x)}
subjected to the equality constraint
g
(
x
)
=
0
{\displaystyle g(x)=0}
, form the Lagrangian function
L
(
x
,
λ
)
=
f
(
x
)
−
λ
g
(
x
)
{\displaystyle {\mathcal {L}}(x,\lambda )=f(x)-\lambda g(x)}
and find the stationary points of
L
{\displaystyle {\mathcal {L}}}
considered as a function of
x
{\displaystyle x}
and the Lagrange multiplier
λ
{\displaystyle \lambda }
. The solution corresponding to the original constrained optimization is always a saddle point of the Lagrangian function, which can be identified among the stationary points from the definiteness of the bordered Hessian matrix.The great advantage of this method is that it allows the optimization to be solved without explicit parameterization in terms of the constraints. As a result, the method of Lagrange multipliers is widely used to solve challenging constrained optimization problems. Further, the method of Lagrange multipliers is generalized by the Karush–Kuhn–Tucker conditions, which can also take into account inequality constraints of the form
h
(
x
)
≤
c
{\displaystyle h(\mathbf {x} )\leq c}
.
View More On Wikipedia.org
Forums
Back
Top