โˆžCalculus IV

Multivariable Calculus Theorems

Study smarter with Fiveable

Get study guides, practice questions, and cheatsheets for all your subjects. Join 500,000+ students with a 96% pass rate.

Get Started

Why This Matters

These theorems aren't just abstract formulas. They're the connective tissue of vector calculus, linking line integrals, surface integrals, volume integrals, and derivatives into a coherent framework. You're being tested on your ability to recognize when to apply each theorem, how they relate to one another, and why certain conditions (like conservative fields or continuous partial derivatives) make computations dramatically simpler. The big picture: these results let you convert hard integrals into easier ones and reveal deep relationships between local behavior (derivatives, divergence, curl) and global behavior (integrals over boundaries).

Don't just memorize the theorem statements. Know what type of problem each theorem solves and how they form a hierarchy. Green's Theorem is a special case of Stokes' Theorem, which connects to the Divergence Theorem through the broader framework of differential forms. When you see an integral, ask yourself: Can I convert this to a simpler domain? Is this field conservative? That conceptual reflex is what separates strong exam performance from mere formula recall.


Fundamental Theorems Connecting Integrals and Derivatives

These theorems share a common structure: they relate an integral over a region to an integral over its boundary, reducing dimensional complexity and revealing when path independence applies.

Gradient Theorem (Fundamental Theorem of Line Integrals)

If F=โˆ‡f\mathbf{F} = \nabla f (meaning F\mathbf{F} is a conservative field with potential function ff), then the line integral depends only on the endpoints:

โˆซCFโ‹…dr=f(b)โˆ’f(a)\int_C \mathbf{F} \cdot d\mathbf{r} = f(\mathbf{b}) - f(\mathbf{a})

  • Path independence is the key payoff. You skip parameterization entirely and just evaluate the potential function at the two endpoints.
  • Conservative field test: If โˆ‡ร—F=0\nabla \times \mathbf{F} = \mathbf{0} on a simply connected domain, then F\mathbf{F} has a potential function. The simply connected condition matters: on domains with holes, a zero curl field can still fail to be conservative.

Green's Theorem

Green's Theorem converts between a line integral around a closed curve CC and a double integral over the enclosed planar region RR. It has two forms:

  • Circulation-curl form: โˆฎCFโ‹…dr=โˆฌR(โˆ‚Qโˆ‚xโˆ’โˆ‚Pโˆ‚y)dA\oint_C \mathbf{F} \cdot d\mathbf{r} = \iint_R \left( \frac{\partial Q}{\partial x} - \frac{\partial P}{\partial y} \right) dA, where F=โŸจP,QโŸฉ\mathbf{F} = \langle P, Q \rangle and CC is traversed counterclockwise (positive orientation).
  • Flux-divergence form: โˆฎCFโ‹…nโ€‰ds=โˆฌR(โˆ‚Pโˆ‚x+โˆ‚Qโˆ‚y)dA\oint_C \mathbf{F} \cdot \mathbf{n} \, ds = \iint_R \left( \frac{\partial P}{\partial x} + \frac{\partial Q}{\partial y} \right) dA, which relates outward flux across CC to the divergence over RR.
  • Area computation trick: Setting P=โˆ’yP = -y and Q=xQ = x gives Area=12โˆฎC(xโ€‰dyโˆ’yโ€‰dx)\text{Area} = \frac{1}{2} \oint_C (x \, dy - y \, dx). This lets you compute enclosed area from a line integral, which is useful when the boundary has a nice parameterization but the region doesn't.

Stokes' Theorem

Stokes' Theorem relates the surface integral of curl over an oriented surface SS to the line integral around its boundary curve C=โˆ‚SC = \partial S:

โˆฌS(โˆ‡ร—F)โ‹…dS=โˆฎCFโ‹…dr\iint_S (\nabla \times \mathbf{F}) \cdot d\mathbf{S} = \oint_C \mathbf{F} \cdot d\mathbf{r}

  • This generalizes Green's Theorem to arbitrary oriented surfaces in R3\mathbb{R}^3. Green's is just the special case where SS is a flat region in the xyxy-plane.
  • Orientation must be consistent: the boundary curve's direction and the surface normal must agree via the right-hand rule. Getting this wrong flips the sign.
  • Curl interpretation: โˆ‡ร—F\nabla \times \mathbf{F} measures local rotation of the field. If โˆ‡ร—F=0\nabla \times \mathbf{F} = \mathbf{0} everywhere on a simply connected domain, the field is conservative.

Divergence Theorem (Gauss's Theorem)

The Divergence Theorem converts a flux integral over a closed surface SS into a volume integral over the enclosed region VV:

โˆญVโˆ‡โ‹…Fโ€‰dV=โˆฏSFโ‹…dS\iiint_V \nabla \cdot \mathbf{F} \, dV = \oiint_S \mathbf{F} \cdot d\mathbf{S}

  • Use this when computing flux directly over SS is painful but the divergence โˆ‡โ‹…F\nabla \cdot \mathbf{F} is simple (e.g., a constant or polynomial).
  • Physical interpretation: divergence measures source/sink strength at a point. The theorem says the net outward flux through the boundary equals the total source strength inside the volume.
  • The surface must be closed (no boundary edges). If it's not closed, you can't apply this theorem directly.

Compare: Green's Theorem vs. Stokes' Theorem: both relate circulation to curl, but Green's is restricted to flat regions in R2\mathbb{R}^2 while Stokes' handles arbitrary surfaces in R3\mathbb{R}^3. If an exam asks you to evaluate a line integral around a curve bounding a surface, Stokes' is your tool.

Compare: Stokes' Theorem vs. Divergence Theorem: Stokes' connects 2D surfaces to 1D boundaries (via curl), while Divergence connects 3D volumes to 2D boundaries (via divergence). Both reduce dimension by one, but they involve different differential operators.


Differentiation Rules for Multivariable Functions

These theorems establish how derivatives behave when functions depend on multiple variables, ensuring consistency and enabling chain-rule computations.

Chain Rule for Multivariable Functions

When you have a composite function where the intermediate variables themselves depend on other variables, the multivariable chain rule tells you how to differentiate through the layers.

For z=f(x,y)z = f(x, y) where x=x(t)x = x(t) and y=y(t)y = y(t):

dzdt=โˆ‚fโˆ‚xdxdt+โˆ‚fโˆ‚ydydt\frac{dz}{dt} = \frac{\partial f}{\partial x}\frac{dx}{dt} + \frac{\partial f}{\partial y}\frac{dy}{dt}

  • Tree diagram method: Draw the dependency structure. Each path from the output variable down to the variable you're differentiating with respect to contributes one term. You multiply along each path and sum all paths.
  • Jacobian generalization: For vector-valued functions, the chain rule becomes matrix multiplication: D(gโˆ˜f)=Dgโ‹…DfD(\mathbf{g} \circ \mathbf{f}) = D\mathbf{g} \cdot D\mathbf{f}, where DfD\mathbf{f} is the Jacobian matrix of f\mathbf{f}.

Clairaut's Theorem (Symmetry of Mixed Partials)

If both mixed partial derivatives โˆ‚2fโˆ‚xโˆ‚y\frac{\partial^2 f}{\partial x \partial y} and โˆ‚2fโˆ‚yโˆ‚x\frac{\partial^2 f}{\partial y \partial x} exist and are continuous near a point, then they're equal at that point:

โˆ‚2fโˆ‚xโˆ‚y=โˆ‚2fโˆ‚yโˆ‚x\frac{\partial^2 f}{\partial x \partial y} = \frac{\partial^2 f}{\partial y \partial x}

  • The continuity requirement is not optional. Counterexamples exist (the classic one involves a piecewise function with f(0,0)=0f(0,0) = 0) where the mixed partials exist but differ because they aren't continuous.
  • Practical use: This simplifies Hessian matrix construction (the Hessian is symmetric for smooth functions) and serves as a quick check on your partial derivative computations.

Compare: The Chain Rule tells you how to differentiate composites, while Clairaut's tells you when differentiation order doesn't matter. Both require smoothness assumptions for validity.


Approximation and Local Behavior

These results let you approximate complicated functions near a point and understand implicit relationships between variables.

Taylor's Theorem for Multivariable Functions

Just as single-variable Taylor series approximate a function near a point using derivatives, the multivariable version uses partial derivatives and the Hessian matrix:

f(x0+h)โ‰ˆf(x0)+โˆ‡f(x0)โ‹…h+12hTH(x0)โ€‰h+โ‹ฏf(\mathbf{x}_0 + \mathbf{h}) \approx f(\mathbf{x}_0) + \nabla f(\mathbf{x}_0) \cdot \mathbf{h} + \frac{1}{2}\mathbf{h}^T H(\mathbf{x}_0) \, \mathbf{h} + \cdots

  • The first-order terms (gradient) give the best linear approximation (the tangent plane). The second-order terms involve the Hessian matrix HH, whose entries are the second partial derivatives Hij=โˆ‚2fโˆ‚xiโˆ‚xjH_{ij} = \frac{\partial^2 f}{\partial x_i \partial x_j}.
  • The quadratic term 12hTHh\frac{1}{2}\mathbf{h}^T H \mathbf{h} is critical for classifying critical points via the second derivative test: the eigenvalues of HH (or its determinant and trace in 2D) tell you whether you have a local min, local max, or saddle point.
  • Error bounds depend on higher-order derivatives over the region, which matters for numerical analysis applications.

Implicit Function Theorem

When you have a constraint equation F(x,y)=0F(x, y) = 0 and you want to know whether you can solve for yy as a function of xx near some point, this theorem gives the answer.

  • Existence guarantee: If F(a,b)=0F(a, b) = 0 and โˆ‚Fโˆ‚y(a,b)โ‰ 0\frac{\partial F}{\partial y}(a, b) \neq 0, then there exists a smooth function y=g(x)y = g(x) defined near x=ax = a satisfying F(x,g(x))=0F(x, g(x)) = 0.
  • Derivative formula: You can find dydx\frac{dy}{dx} without ever solving for yy explicitly:

dydx=โˆ’โˆ‚F/โˆ‚xโˆ‚F/โˆ‚y\frac{dy}{dx} = -\frac{\partial F/\partial x}{\partial F/\partial y}

  • Higher dimensions: For a system F(x,y)=0\mathbf{F}(\mathbf{x}, \mathbf{y}) = \mathbf{0}, the condition becomes that the Jacobian matrix of F\mathbf{F} with respect to the dependent variables y\mathbf{y} must be non-singular (invertible) at the point.

Compare: Taylor's Theorem approximates a known function locally, while the Implicit Function Theorem guarantees a function exists from a constraint equation. Both are local results requiring smoothness.


Optimization Under Constraints

This theorem provides the standard method for constrained optimization, appearing constantly in applications and exams.

Lagrange Multiplier Theorem

When you need to optimize f(x)f(\mathbf{x}) subject to a constraint g(x)=cg(\mathbf{x}) = c, the method of Lagrange multipliers says that at any constrained extremum (where โˆ‡gโ‰ 0\nabla g \neq \mathbf{0}):

โˆ‡f=ฮปโˆ‡g\nabla f = \lambda \nabla g

  • Geometric interpretation: At the optimal point, the level curves of ff are tangent to the constraint surface. If they weren't tangent, you could move along the constraint and still increase (or decrease) ff.
  • Solving the system: You get n+1n + 1 equations (the nn component equations from โˆ‡f=ฮปโˆ‡g\nabla f = \lambda \nabla g plus the constraint g(x)=cg(\mathbf{x}) = c) in n+1n + 1 unknowns (the nn coordinates plus ฮป\lambda).
  • Multiple constraints: With kk constraints, use kk multipliers: โˆ‡f=ฮป1โˆ‡g1+ฮป2โˆ‡g2+โ‹ฏ+ฮปkโˆ‡gk\nabla f = \lambda_1 \nabla g_1 + \lambda_2 \nabla g_2 + \cdots + \lambda_k \nabla g_k.
  • The multiplier ฮป\lambda itself has meaning: it measures the sensitivity of the optimal value to small changes in the constraint constant cc.

Compare: Without constraints, you set โˆ‡f=0\nabla f = \mathbf{0} directly. With constraints, the gradient of ff doesn't need to be zero; it just needs to be a linear combination of the constraint gradients. That's the core shift in thinking.


Quick Reference Table

ConceptBest Examples
Boundary-integral relationshipsGradient Theorem, Green's, Stokes', Divergence Theorem
Path independenceGradient Theorem, conservative field tests
2D region โ†” boundaryGreen's Theorem
3D surface โ†” boundary curveStokes' Theorem
3D volume โ†” boundary surfaceDivergence Theorem
Differentiation mechanicsChain Rule, Clairaut's Theorem
Local approximationTaylor's Theorem, Implicit Function Theorem
Constrained optimizationLagrange Multiplier Theorem

Self-Check Questions

  1. Which two theorems both relate circulation around a boundary to a "curl-type" integral, and what distinguishes their domains of application?

  2. If you're given a vector field and asked whether a line integral is path-independent, which theorem justifies your answer, and what condition must you verify?

  3. Compare the Divergence Theorem and Stokes' Theorem: what type of integral does each convert, and what differential operator appears in each?

  4. An exam gives you F(x,y,z)=0F(x, y, z) = 0 and asks for โˆ‚zโˆ‚x\frac{\partial z}{\partial x}. Which theorem applies, and what must be nonzero for it to work?

  5. You need to maximize f(x,y)f(x, y) subject to g(x,y)=kg(x, y) = k. Write the system of equations you must solve, and explain geometrically why โˆ‡f\nabla f and โˆ‡g\nabla g must be parallel at the solution.

Multivariable Calculus Theorems to Know for Calculus IV