Study smarter with Fiveable
Get study guides, practice questions, and cheatsheets for all your subjects. Join 500,000+ students with a 96% pass rate.
Lagrange multipliers give you a systematic way to optimize a function when constraints are present. You can't spend infinite money, use unlimited materials, or ignore physical boundaries, so nearly every real optimization problem involves some restriction. This method handles those restrictions directly, and it shows up again in differential equations, physics, and machine learning.
For your exam, algebraic manipulation alone won't cut it. You need to understand why gradients must be parallel at optimal points, how the multiplier encodes sensitivity information, and when this method applies versus other optimization approaches. That conceptual layer is what gets tested hardest on free-response problems.
Every Lagrange multiplier problem starts with constructing the right function. This transforms a constrained problem into one where standard calculus techniques apply.
The Lagrangian combines your objective function with the constraint through the multiplier :
Setting generates your full system of equations. This single expression encapsulates both the optimization goal and the constraint requirement. The constraint itself appears as one of the solving equations, which ensures any critical point you find actually lives on the constraint.
Compare: The Lagrangian setup vs. direct substitution. Both solve constrained problems, but Lagrange multipliers preserve the constraint's identity and give you , which carries valuable sensitivity information. If a problem asks about "rate of change with respect to the constraint," you need the Lagrange approach.
Understanding the geometric intuition turns Lagrange multipliers from a memorized procedure into something that feels inevitable. The core idea: optimal points occur where level curves of are tangent to the constraint curve.
At an optimal point, the gradient of and the gradient of must be parallel:
This parallel condition, combined with the constraint , generates equations for variables plus . The sign of isn't arbitrary either: it reflects whether the constraint is "pushing" the optimum up or down.
Picture the level curves of as contour lines on a map, and the constraint as a trail you must stay on. The optimum occurs where a level curve just touches the constraint without crossing it.
Why tangency? Because is perpendicular to level curves of , and is perpendicular to the constraint curve. If these two normals are parallel, the curves are tangent. At any non-tangent intersection, you could slide along the constraint and reach a higher (or lower) level curve, so that point can't be an extremum.
Compare: Geometric interpretation vs. algebraic setup. The algebra gives you equations to solve, but the geometry explains why those equations work. Exam questions often ask you to sketch or explain why gradients must be parallel.
Setting up the Lagrangian is only half the battle. The system of equations you generate requires careful algebraic technique and verification.
Here's the standard workflow:
Write out the partial derivative equations:
This gives you three equations in three unknowns: , , and .
Eliminate strategically. Often the cleanest move is to divide the first two equations (when possible) to cancel , reducing the system to a relationship between and .
Substitute back into the constraint to solve for the remaining variables.
Verify your critical points. Each one is only a candidate. You must determine whether it's a maximum, minimum, or neither using second-order tests (bordered Hessian) or by comparing function values at all critical points.
Compare: Solving Lagrange systems vs. solving unconstrained critical points. Both require setting partial derivatives to zero, but Lagrange problems have the extra equation from the constraint and the extra unknown . Budget your time accordingly on exams.
The Lagrange multiplier isn't just a computational artifact. It carries real economic and physical meaning, and examiners test this regularly.
The multiplier measures sensitivity of the optimal value to changes in the constraint:
If you relax the constraint by a small amount , the optimal value of changes by approximately . In economics, this is called the shadow price: it tells you how much an additional unit of the constrained resource is worth in terms of your objective.
The sign matters too. Positive means increasing increases the optimum; negative means the opposite.
Compare: in utility maximization vs. cost minimization. In both cases measures marginal value, but the interpretation flips: one tells you the value of more budget, the other tells you the cost of more output. Free-response questions often ask you to interpret in context.
Real problems often involve several constraints simultaneously, and knowing when Lagrange multipliers are the right tool matters.
For constraints, add one multiplier per constraint:
Each tracks the marginal impact of its corresponding constraint. The system grows accordingly: variables plus multipliers gives you equations. This stays manageable for small systems but gets computationally heavy as and increase.
Geometrically, you're finding where the objective's level surface is tangent to the intersection of all constraint surfaces. The gradient of must lie in the span of the constraint gradients .
| Method | Best For | Limitation |
|---|---|---|
| Lagrange multipliers | Equality constraints () | Can't handle inequalities directly |
| KKT conditions | Inequality constraints () | More complex; requires complementary slackness |
| Direct substitution | Simple constraints you can solve for one variable | Loses sensitivity info (no ) |
| Gradient descent | Unconstrained optimization | Requires modification (projected gradient) for constraints |
The key distinction for your exam: if the problem says "subject to " with an equals sign, use Lagrange. If it's an inequality , you'd need Karush-Kuhn-Tucker (KKT) conditions, which generalize Lagrange multipliers by adding complementary slackness conditions.
Compare: Lagrange multipliers vs. KKT conditions. Lagrange handles (equality); KKT handles (inequality). If the constraint is an inequality, first check whether it's binding (active as equality) at the solution. If it is, Lagrange still works for that case.
| Concept | Key Details |
|---|---|
| Lagrangian setup | ; set all partials to zero |
| Gradient condition | at optimal points |
| Geometric meaning | Level curve of tangent to constraint curve |
| Interpretation of | Sensitivity of optimum to constraint relaxation |
| Multiple constraints | Add for each |
| Constraint qualification | Requires at solution points |
| Economic applications | Utility maximization, cost minimization, shadow prices |
| Method limitations | Equality constraints only; KKT needed for inequalities |
If and point in opposite directions at a critical point, what does this tell you about the sign of , and what does that sign mean economically?
Compare solving a constrained optimization problem using direct substitution versus Lagrange multipliers. What information do you gain from the Lagrange approach that substitution doesn't provide?
You find two critical points from a Lagrange system. How do you determine which gives the maximum and which gives the minimum of the objective function?
In a utility maximization problem with budget constraint , you calculate . Explain in plain language what this tells the consumer.
When would you need to extend the basic Lagrange multiplier method to KKT conditions, and what's the key difference in the type of constraint each handles?