Why This Matters
Partial derivatives are your gateway to understanding how multivariable functions behave. In Calculus II, you need to analyze functions that depend on more than one variable at a time. This isn't just an extension of single-variable calculus; it's a fundamentally different way of thinking about change. You'll need to master how functions respond when you tweak one variable while holding others fixed, how the gradient captures the direction of steepest increase, and how optimization works when you have multiple inputs.
These concepts connect directly to exam applications: finding tangent planes, optimizing functions with constraints, and understanding how composite functions behave through the multivariable chain rule. Don't just memorize formulas. Know why we treat other variables as constants, when Clairaut's theorem applies, and how the gradient relates to directional derivatives.
Foundations: What Partial Derivatives Measure
A partial derivative isolates the effect of one variable on a function's output while treating all other variables as constants. Think of it this way: if you have a surface in 3D, a partial derivative tells you the slope of that surface along one coordinate axis at a time.
Definition of Partial Derivatives
- Measures the rate of change of a function with respect to one variable while all other variables remain fixed. Geometrically, you're slicing a 3D surface with a vertical plane parallel to one axis and finding the slope of the resulting curve.
- Conceptually distinct from ordinary derivatives. You're examining change along one axis only, not total change across all variables simultaneously.
Notation for Partial Derivatives
The symbol โ distinguishes partial derivatives from ordinary derivatives (d). You'll see three common notations used interchangeably:
- Leibniz notation: โxโfโ
- Subscript notation: fxโ or โxโf
- Higher-order notation uses repeated symbols: โx2โ2fโ (or fxxโ) for second partials, and โxโyโ2fโ (or fxyโ) for mixed partials
Subscript notation is faster for computation. fxxโ means differentiate twice with respect to x, while fxyโ means differentiate first with respect to x, then with respect to y.
Calculating Partial Derivatives
Here's the process:
- Identify your active variable before you start differentiating. This prevents the most common careless errors.
- Treat all other variables as constants. For example, to find โxโโ of f(x,y)=x2y+y3, the y in the first term acts like a coefficient and the y3 term is just a constant. The result is 2xy+0=2xy.
- Apply standard differentiation rules (product, quotient, chain) exactly as in single-variable calculus. The only difference is what counts as a constant.
Compare: Partial derivatives vs. ordinary derivatives: both measure instantaneous rate of change, but partial derivatives hold other variables fixed while ordinary derivatives allow all variables to change. When writing solutions, always state explicitly which variables you're treating as constants.
Higher-Order Behavior: Second Derivatives and Beyond
Higher-order partial derivatives reveal curvature and concavity of multivariable functions. They're essential for optimization and understanding surface geometry.
Higher-Order Partial Derivatives
- Second partial derivatives like fxxโ and fyyโ measure how the rate of change itself changes. This is analogous to concavity in single-variable calculus: fxxโ tells you about curvature along the x-direction.
- These are central to the second derivative test for classifying critical points as maxima, minima, or saddle points (covered in the optimization section below).
Mixed Partial Derivatives
- Differentiate with respect to different variables in succession. fxyโ means: first compute fxโ, then differentiate that result with respect to y.
- A nonzero mixed partial indicates variable interaction. It tells you that changing one variable affects how the function responds to changes in the other.
- Watch the order of notation. In Leibniz form, โyโxโ2fโ means differentiate with respect to x first, then y (read right to left). In subscript form, fxyโ means differentiate with respect to x first, then y (read left to right).
Clairaut's Theorem (Equality of Mixed Partials)
Clairaut's theorem states that fxyโ=fyxโ provided both mixed partials are continuous near the point in question. For virtually every function you'll encounter in this course, this condition holds.
Why does this matter? It simplifies calculations dramatically. If one order of differentiation involves simpler algebra than the other, you can choose the easier path and get the same answer.
Compare: fxyโ vs. fyxโ: by Clairaut's theorem, these are equal for functions with continuous second partials, but the computation path can differ. If one order is simpler, use it. This is a common exam shortcut.
The Chain Rule and Implicit Differentiation
When functions are composed or defined implicitly, you need specialized techniques. The multivariable chain rule tracks how changes propagate through nested dependencies.
The Chain Rule for Partial Derivatives
If z=f(x,y) where x=g(t) and y=h(t), then:
dtdzโ=โxโfโdtdxโ+โyโfโdtdyโ
Each term accounts for one pathway through which t affects z. Tree diagrams are genuinely helpful here: draw arrows from the final output back through intermediate variables to the independent variables. Each path through the tree contributes one term to the chain rule expression.
This generalizes naturally. If x and y each depend on two variables s and t, you get partial derivatives of z with respect to s and t, each with two terms summing over the intermediate variables.
Implicit Differentiation with Partial Derivatives
When you can't solve explicitly for one variable, implicit differentiation handles it. Given a surface defined by F(x,y,z)=0, you can find โxโzโ without ever isolating z.
The shortcut formula is:
โxโzโ=โFzโFxโโ(whenย Fzโ๎ =0)
This comes from differentiating F(x,y,z)=0 with respect to x, treating y as constant and z as a function of x and y. Every term containing z picks up a factor of โxโzโ via the chain rule. Then you solve for โxโzโ.
Compare: Explicit vs. implicit differentiation: explicit is straightforward substitution, while implicit handles equations you can't solve for a single variable. Exam problems often give surfaces in implicit form, so the formula โxโzโ=โFzโFxโโ is worth memorizing.
Gradients and Directional Derivatives
The gradient unifies all first-order partial derivatives into a single vector that points toward steepest ascent. Directional derivatives generalize partial derivatives to arbitrary directions.
Gradient Vector
- Defined as โf=โจfxโ,fyโโฉ in 2D (or โจfxโ,fyโ,fzโโฉ in 3D). It collects all first partial derivatives into one vector.
- Points in the direction of steepest increase of f. This is the single most important geometric fact about gradients.
- Its magnitude โฅโfโฅ gives the maximum rate of change of the function at that point.
- Perpendicular to level curves (in 2D) and level surfaces (in 3D). This is why the gradient appears in tangent plane and normal line calculations.
Directional Derivatives
The directional derivative measures the rate of change of f in any direction, not just along coordinate axes.
Duโf=โfโ
u
where u is a unit vector in the desired direction. A common mistake is forgetting to normalize: if you're given a direction vector v, divide by its magnitude to get u=โฅvโฅvโ before computing the dot product.
- Maximum value of Duโf is โฅโfโฅ, achieved when u points in the gradient direction.
- Minimum value is โโฅโfโฅ, achieved in the opposite direction.
- Zero directional derivative occurs when u is perpendicular to โf (moving along a level curve).
Compare: Partial derivatives vs. directional derivatives: partial derivatives are special cases of directional derivatives along coordinate axes (i or j). If asked for the rate of change in a non-axis direction, you need the directional derivative formula.
Geometric Applications: Tangent Planes and Linearization
Partial derivatives let you construct linear approximations to surfaces. This is the multivariable analog of tangent lines.
Tangent Planes and Normal Lines
For a surface z=f(x,y) at the point (a,b,f(a,b)), the tangent plane is:
zโf(a,b)=fxโ(a,b)(xโa)+fyโ(a,b)(yโb)
The normal vector to this plane is โจfxโ(a,b),fyโ(a,b),โ1โฉ. For an implicit surface F(x,y,z)=0, the normal vector is simply โF.
The normal line passes through the point in the direction of the normal vector. You can write it parametrically using the normal vector as the direction.
Taylor Series for Multivariable Functions
Multivariable Taylor series extend polynomial approximation to functions of several variables.
- First-order approximation: f(x,y)โf(a,b)+fxโ(a,b)(xโa)+fyโ(a,b)(yโb). Notice this is exactly the tangent plane equation.
- Second-order terms bring in fxxโ, fyyโ, and fxyโ, giving a better approximation near (a,b). These same second-order terms appear in the discriminant used for the second derivative test.
Compare: Tangent plane vs. linearization: these are the same object with different names. "Tangent plane" emphasizes geometry; "linearization" emphasizes approximation. Both use the same formula.
Optimization and Applications
Partial derivatives are the primary tool for finding extrema of multivariable functions.
Partial Derivatives in Optimization Problems
Finding and classifying critical points follows a clear procedure:
- Find critical points by setting โf=0. This means solving fxโ=0 and fyโ=0 simultaneously.
- Compute the discriminant at each critical point: D=fxxโfyyโโ(fxyโ)2
- Classify using D:
- D>0 and fxxโ>0: local minimum
- D>0 and fxxโ<0: local maximum
- D<0: saddle point
- D=0: test is inconclusive
- Check boundaries if the domain is a constrained region. Critical points alone aren't enough; extrema can also occur on the boundary.
Applications in Physics and Engineering
- Heat equation and diffusion use โtโuโ and โ2u (the Laplacian) to model how temperature distributes over time.
- Fluid dynamics relies on partial derivatives to express conservation laws and flow rates.
- Economics uses partial derivatives for marginal analysis. For instance, โxโCโ gives the marginal cost with respect to one production input while holding others fixed.
Compare: Single-variable vs. multivariable optimization: single-variable uses fโฒ(x)=0 and checks the sign of fโฒโฒ(x). Multivariable uses โf=0 and the discriminant D. The discriminant has no single-variable analog because saddle points don't exist in 1D.
Quick Reference Table
|
| Basic computation | Treat other variables as constants, apply standard rules |
| Higher-order derivatives | fxxโ, fyyโ, fxyโ for curvature and the second derivative test |
| Clairaut's theorem | fxyโ=fyxโ when mixed partials are continuous |
| Chain rule | dtdzโ=โxโfโdtdxโ+โyโfโdtdyโ |
| Implicit differentiation | โxโzโ=โFzโFxโโ for implicit surfaces |
| Gradient | โf=โจfxโ,fyโโฉ, direction of steepest ascent |
| Directional derivative | Duโf=โfโ
u, rate of change in direction u |
| Tangent plane | zโz0โ=fxโ(a,b)(xโa)+fyโ(a,b)(yโb) |
| Optimization | Critical points where โf=0, discriminant D=fxxโfyyโโfxy2โ |
Self-Check Questions
-
If f(x,y)=x3y2+sin(xy), what is fxyโ? Verify that fxyโ=fyxโ and explain why Clairaut's theorem guarantees this.
-
Compare the gradient vector and the directional derivative: how does knowing โf at a point let you find the rate of change in any direction?
-
For the surface x2+y2+z2=9, use implicit differentiation to find โxโzโ. What does this derivative represent geometrically?
-
A critical point has fxxโ=4, fyyโ=3, and fxyโ=5. Calculate the discriminant D and classify the critical point. What would change if fxyโ=2 instead?
-
Contrast finding the tangent plane to z=f(x,y) at a point versus finding the tangent plane to an implicitly defined surface F(x,y,z)=0. What role does the gradient play in each case?