โˆžCalculus IV

Partial Derivative Rules

Study smarter with Fiveable

Get study guides, practice questions, and cheatsheets for all your subjects. Join 500,000+ students with a 96% pass rate.

Get Started

Why This Matters

Partial derivatives are the foundation of Calculus IV. They extend differentiation to functions with multiple inputs. Whenever you analyze a surface, optimize a multivariable function, or study how a physical system responds to changes, you're relying on partial derivatives. The rules here connect directly to gradient vectors, directional derivatives, optimization, and vector field analysis.

You'll be tested on more than computation. Examiners want to see that you understand when to apply each rule and why it works. The chain rule for partials drives related rates problems. Clairaut's theorem saves you time on mixed partials. The gradient ties everything together for optimization. Knowing what concept each rule captures, and when to reach for it, matters just as much as cranking through algebra.


Foundational Concepts and Notation

Before applying any rules, you need a solid understanding of what partial derivatives actually measure and how to communicate them precisely. A partial derivative isolates the rate of change in one direction while treating all other variables as constants.

Definition of Partial Derivatives

The formal definition mirrors the single-variable limit definition. For a function f(x,y)f(x, y):

fx(a,b)=limโกhโ†’0f(a+h,โ€‰b)โˆ’f(a,โ€‰b)hf_x(a, b) = \lim_{h \to 0} \frac{f(a+h,\, b) - f(a,\, b)}{h}

  • Measures single-variable change: a partial derivative tells you how ff changes as one variable varies while all others stay fixed.
  • Enables independent analysis of each variable's contribution to the function's behavior.
  • Notation โˆ‚fโˆ‚x\frac{\partial f}{\partial x} uses the curly โˆ‚\partial to distinguish partial differentiation from total (ordinary) differentiation.

Partial Derivative Notation

  • First-order partials use โˆ‚fโˆ‚x\frac{\partial f}{\partial x}, โˆ‚fโˆ‚y\frac{\partial f}{\partial y}, or subscript notation like fxf_x, fyf_y.
  • Mixed second derivatives written as โˆ‚2fโˆ‚xโˆ‚y\frac{\partial^2 f}{\partial x \partial y} mean: differentiate first with respect to yy, then with respect to xx. Read right to left in Leibniz notation.
  • Higher-order notation extends naturally: โˆ‚3fโˆ‚x2โˆ‚y\frac{\partial^3 f}{\partial x^2 \partial y} means differentiate once with respect to yy, then twice with respect to xx.

Compare: โˆ‚fโˆ‚x\frac{\partial f}{\partial x} vs. dfdx\frac{df}{dx}. Both use Leibniz-style notation, but the curly โˆ‚\partial signals a multivariable context where other variables are held constant. If a problem uses โˆ‚\partial, you're in partial derivative territory.


Differentiation Rules for Partial Derivatives

The computational rules from single-variable calculus (product rule, quotient rule, chain rule) all extend to partial derivatives with one key modification: treat every variable except the one you're differentiating with respect to as a constant.

Partial Derivatives of Multivariable Functions

Here's the process:

  1. Identify which variable you're differentiating with respect to.
  2. Treat all other variables as constants.
  3. Apply the usual single-variable differentiation rules (power, product, quotient, exponential, trig, etc.).

Example: For f(x,y)=x3y2+sinโก(xy)f(x, y) = x^3 y^2 + \sin(xy), finding fxf_x:

  • Treat yy as a constant.
  • โˆ‚โˆ‚x(x3y2)=3x2y2\frac{\partial}{\partial x}(x^3 y^2) = 3x^2 y^2 (power rule; y2y^2 is just a constant coefficient).
  • โˆ‚โˆ‚x(sinโก(xy))=ycosโก(xy)\frac{\partial}{\partial x}(\sin(xy)) = y\cos(xy) (chain rule; derivative of the inner function xyxy with respect to xx is yy).
  • So fx=3x2y2+ycosโก(xy)f_x = 3x^2 y^2 + y\cos(xy).

Setting all first partials equal to zero simultaneously is how you find critical points of multivariable functions.

Chain Rule for Partial Derivatives

This is where things get genuinely tricky. The chain rule has different forms depending on the dependency structure.

Case 1: Single parameter. If z=f(x,y)z = f(x, y) where x=x(t)x = x(t) and y=y(t)y = y(t), then:

dzdt=โˆ‚fโˆ‚xdxdt+โˆ‚fโˆ‚ydydt\frac{dz}{dt} = \frac{\partial f}{\partial x}\frac{dx}{dt} + \frac{\partial f}{\partial y}\frac{dy}{dt}

Notice the result is an ordinary derivative dzdt\frac{dz}{dt} because zz ultimately depends on only one independent variable.

Case 2: Multiple parameters. If z=f(x,y)z = f(x, y) where x=x(s,t)x = x(s, t) and y=y(s,t)y = y(s, t), then:

โˆ‚zโˆ‚s=โˆ‚fโˆ‚xโˆ‚xโˆ‚s+โˆ‚fโˆ‚yโˆ‚yโˆ‚s\frac{\partial z}{\partial s} = \frac{\partial f}{\partial x}\frac{\partial x}{\partial s} + \frac{\partial f}{\partial y}\frac{\partial y}{\partial s}

and similarly for โˆ‚zโˆ‚t\frac{\partial z}{\partial t}.

Tree diagrams are genuinely useful here. Draw zz at the top, connect it to xx and yy, then connect those to ss and tt. Each path from zz down to your target variable contributes one term (multiply along branches, add across paths).

Implicit Differentiation for Multivariable Functions

When a relationship is given as F(x,y,z)=0F(x, y, z) = 0 and you can't easily solve for zz, implicit differentiation gives you the partial derivatives directly.

Formula: If F(x,y,z)=0F(x, y, z) = 0 and Fzโ‰ 0F_z \neq 0, then:

โˆ‚zโˆ‚x=โˆ’FxFzandโˆ‚zโˆ‚y=โˆ’FyFz\frac{\partial z}{\partial x} = -\frac{F_x}{F_z} \qquad \text{and} \qquad \frac{\partial z}{\partial y} = -\frac{F_y}{F_z}

Why this works: Differentiate F(x,y,z)=0F(x, y, z) = 0 with respect to xx using the chain rule (treating zz as a function of xx and yy), then solve for โˆ‚zโˆ‚x\frac{\partial z}{\partial x}.

Compare: Standard partial differentiation vs. implicit differentiation. Both find partial derivatives, but implicit differentiation works when you can't isolate the dependent variable. Use implicit when you see equations like x2+y2+z2=1x^2 + y^2 + z^2 = 1 rather than z=1โˆ’x2โˆ’y2z = \sqrt{1 - x^2 - y^2}.


Symmetry and Higher-Order Derivatives

When you take multiple partial derivatives, the order can matter. Clairaut's theorem tells you exactly when you can swap the order, which is most of the time for functions you'll encounter in this course.

Clairaut's Theorem (Equality of Mixed Partials)

Statement: If fxyf_{xy} and fyxf_{yx} are both continuous on an open region, then:

โˆ‚2fโˆ‚xโˆ‚y=โˆ‚2fโˆ‚yโˆ‚x\frac{\partial^2 f}{\partial x \partial y} = \frac{\partial^2 f}{\partial y \partial x}

  • Computation shortcut: choose whichever differentiation order is easier. The result is identical as long as the continuity condition holds.
  • The continuity requirement is almost always satisfied for functions on exams. The exception to watch for is piecewise-defined functions at the boundary point, where the mixed partials can fail to be continuous and may differ.

Higher-Order Partial Derivatives

  • Pure second partials like fxx=โˆ‚2fโˆ‚x2f_{xx} = \frac{\partial^2 f}{\partial x^2} measure concavity in the xx-direction, just like single-variable second derivatives.
  • Mixed partials like fxyf_{xy} capture how the rate of change in one direction varies as you move in another direction. Think of it as measuring "twist."
  • The Hessian matrix collects all second partials into a matrix:

H=[fxxfxyfxyfyy]H = \begin{bmatrix} f_{xx} & f_{xy} \\ f_{xy} & f_{yy} \end{bmatrix}

The second derivative test uses the Hessian's determinant: D=fxxfyyโˆ’(fxy)2D = f_{xx}f_{yy} - (f_{xy})^2. At a critical point where โˆ‡f=0\nabla f = \mathbf{0}:

  • D>0D > 0 and fxx>0f_{xx} > 0: local minimum
  • D>0D > 0 and fxx<0f_{xx} < 0: local maximum
  • D<0D < 0: saddle point
  • D=0D = 0: test is inconclusive

Compare: โˆ‚2fโˆ‚x2\frac{\partial^2 f}{\partial x^2} vs. โˆ‚2fโˆ‚xโˆ‚y\frac{\partial^2 f}{\partial x \partial y}. Pure second partials measure curvature along axes, while mixed partials measure twist. Both appear in the discriminant D=fxxfyyโˆ’(fxy)2D = f_{xx}f_{yy} - (f_{xy})^2.


Gradient and Directional Analysis

The gradient packages all first partial derivatives into a single vector, unlocking powerful geometric interpretations. It points toward steepest increase, and its magnitude tells you how steep.

Gradient Vector

Definition: For f(x,y,z)f(x, y, z):

โˆ‡f=โŸจโˆ‚fโˆ‚x,โ€‰โˆ‚fโˆ‚y,โ€‰โˆ‚fโˆ‚zโŸฉ=โŸจfx,โ€‰fy,โ€‰fzโŸฉ\nabla f = \left\langle \frac{\partial f}{\partial x},\, \frac{\partial f}{\partial y},\, \frac{\partial f}{\partial z} \right\rangle = \langle f_x,\, f_y,\, f_z \rangle

  • Points toward steepest ascent: moving in the direction of โˆ‡f\nabla f increases ff as rapidly as possible.
  • Magnitude โˆฅโˆ‡fโˆฅ\|\nabla f\| equals the maximum rate of change of ff at that point.
  • Normal to level curves/surfaces: โˆ‡f\nabla f at a point is perpendicular to the level curve (2D) or level surface (3D) passing through that point. This is why the gradient shows up in tangent plane equations.

Directional Derivatives

The directional derivative gives the rate of change of ff in any direction, not just along coordinate axes.

Duf=โˆ‡fโ‹…uD_{\mathbf{u}}f = \nabla f \cdot \mathbf{u}

where u\mathbf{u} must be a unit vector (โˆฅuโˆฅ=1\|\mathbf{u}\| = 1). If you're given a direction vector that isn't unit length, normalize it first.

Three facts worth memorizing:

  • Maximum Duf=โˆฅโˆ‡fโˆฅD_{\mathbf{u}}f = \|\nabla f\|, occurring when \mathbf{u}} points in the gradient direction.
  • Minimum Duf=โˆ’โˆฅโˆ‡fโˆฅD_{\mathbf{u}}f = -\|\nabla f\|, occurring in the direction opposite the gradient.
  • Zero directional derivative occurs when u\mathbf{u} is perpendicular to โˆ‡f\nabla f, meaning you're moving along a level curve/surface.

Compare: Gradient vs. directional derivative. The gradient gives you the direction of maximum increase, while the directional derivative gives you the rate of change in any specified direction. A common exam question: "In what direction does ff increase most rapidly?" Answer: the gradient direction.


Vector-Valued Extensions

When your function outputs a vector instead of a scalar, partial derivatives apply component-by-component. This extends naturally to analyzing vector fields and parametric surfaces.

Partial Derivatives of Vector-Valued Functions

For a parametric surface r(u,v)=โŸจx(u,v),โ€‰y(u,v),โ€‰z(u,v)โŸฉ\mathbf{r}(u, v) = \langle x(u,v),\, y(u,v),\, z(u,v) \rangle, take partials of each component separately:

ru=โŸจโˆ‚xโˆ‚u,โ€‰โˆ‚yโˆ‚u,โ€‰โˆ‚zโˆ‚uโŸฉ\mathbf{r}_u = \left\langle \frac{\partial x}{\partial u},\, \frac{\partial y}{\partial u},\, \frac{\partial z}{\partial u} \right\rangle

  • The vectors ru\mathbf{r}_u and rv\mathbf{r}_v are tangent to the surface at each point.
  • Their cross product ruร—rv\mathbf{r}_u \times \mathbf{r}_v gives a normal vector to the surface, which you'll need for surface integrals and flux calculations.

Quick Reference Table

ConceptKey Formula / Idea
Basic computationโˆ‚fโˆ‚x\frac{\partial f}{\partial x}, โˆ‚fโˆ‚y\frac{\partial f}{\partial y} with other variables held constant
Chain ruledzdt=fxdxdt+fydydt\frac{dz}{dt} = f_x \frac{dx}{dt} + f_y \frac{dy}{dt}; use tree diagrams for complex dependencies
Implicit differentiationโˆ‚zโˆ‚x=โˆ’FxFz\frac{\partial z}{\partial x} = -\frac{F_x}{F_z} when F(x,y,z)=0F(x,y,z) = 0
Clairaut's theoremfxy=fyxf_{xy} = f_{yx} when mixed partials are continuous
Second derivative testD=fxxfyyโˆ’(fxy)2D = f_{xx}f_{yy} - (f_{xy})^2; check sign of DD and fxxf_{xx}
Gradient vectorโˆ‡f=โŸจfx,fy,fzโŸฉ\nabla f = \langle f_x, f_y, f_z \rangle; points toward steepest ascent; normal to level surfaces
Directional derivativesDuf=โˆ‡fโ‹…uD_{\mathbf{u}}f = \nabla f \cdot \mathbf{u} where u\mathbf{u} is a unit vector

Self-Check Questions

  1. If f(x,y)=x2y+exyf(x, y) = x^2 y + e^{xy}, which rule do you use to find โˆ‚fโˆ‚x\frac{\partial f}{\partial x}, and what do you treat yy as during the calculation?

  2. Compare โˆ‚2fโˆ‚xโˆ‚y\frac{\partial^2 f}{\partial x \partial y} and โˆ‚2fโˆ‚yโˆ‚x\frac{\partial^2 f}{\partial y \partial x}. Under what condition are they equal, and why does this matter computationally?

  3. Given โˆ‡f=โŸจ3,โˆ’4โŸฉ\nabla f = \langle 3, -4 \rangle at a point, what is the maximum rate of change of ff, and in what direction does it occur?

  4. When would you choose implicit differentiation over standard partial differentiation? Give an example equation where implicit differentiation is the better approach.

  5. FRQ-style: A surface is defined by z=f(x,y)z = f(x, y). Explain how you would use the gradient to find a vector normal to the level curve f(x,y)=cf(x, y) = c and a vector normal to the surface itself. What's the relationship between these two normals?

Partial Derivative Rules to Know for Calculus IV