Why This Matters
Mathematical inequalities aren't just abstract formulas to memorize—they're the workhorses that let mathematicians establish bounds, prove theorems, and solve optimization problems. When you're working through calculus, linear algebra, or probability, you'll constantly need to show that one quantity is larger or smaller than another. These inequalities give you the tools to do exactly that. You're being tested on your ability to recognize when to apply each inequality, understand the conditions required for each to hold, and connect them to broader mathematical structures like norms, convexity, and expected values.
Think of inequalities as a toolkit: some handle geometric relationships, others tame products and sums, and still others give you control over random variables. The key is understanding which tool fits which job. Don't just memorize the formulas—know what mathematical principle each inequality captures and when its conditions are satisfied. That conceptual understanding will serve you far better on exams than rote recall.
Geometric and Metric Inequalities
These inequalities establish fundamental properties of distance and magnitude. They're the foundation for understanding how distances behave in mathematical spaces.
Triangle Inequality
- Absolute values satisfy ∣a+b∣≤∣a∣+∣b∣—the "length" of a sum never exceeds the sum of lengths
- Geometric interpretation: in any triangle, the sum of two side lengths must exceed the third side
- Foundation for metrics—any valid distance function must satisfy this property, making it essential for analysis
Minkowski's Inequality
- Generalizes the triangle inequality to Lp spaces: ∣∣x+y∣∣p≤∣∣x∣∣p+∣∣y∣∣p
- Works for any p≥1, where the p-norm is defined as ∣∣x∣∣p=(∑∣xi∣p)1/p
- Proves that Lp spaces are valid vector spaces—without this inequality, we couldn't define proper distances in these function spaces
Compare: Triangle Inequality vs. Minkowski's Inequality—both establish that "distance to a sum ≤ sum of distances," but Minkowski generalizes this to arbitrary p-norms. If a problem involves Lp spaces or weighted norms, reach for Minkowski; for basic absolute values, the triangle inequality suffices.
Product and Sum Bounding Inequalities
These inequalities help you control products by relating them to sums—crucial when you need to bound a complicated product using simpler terms.
Cauchy-Schwarz Inequality
- The workhorse of linear algebra: (a1b1+⋯+anbn)2≤(a12+⋯+an2)(b12+⋯+bn2)
- Equivalently written using inner products as ∣⟨u,v⟩∣2≤⟨u,u⟩⋅⟨v,v⟩
- Equality holds exactly when vectors are parallel—use this condition to identify when bounds are tight
Hölder's Inequality
- Generalizes Cauchy-Schwarz to conjugate exponents: ∑∣aibi∣≤(∑∣ai∣p)1/p(∑∣bi∣q)1/q where p1+q1=1
- Cauchy-Schwarz is the special case p=q=2—recognizing this connection helps you choose the right tool
- Essential for functional analysis and proving integrability results in Lp spaces
Young's Inequality
- Bounds products by sums of powers: ab≤pap+qbq for non-negative a,b and conjugates p1+q1=1
- Often used to prove Hölder's inequality—it's lower-level in the hierarchy of tools
- Key technique: when you see a product you need to bound, try splitting it using Young's with strategic exponent choices
Compare: Cauchy-Schwarz vs. Hölder's—Cauchy-Schwarz uses the L2 norm (sum of squares), while Hölder's works for any conjugate pair (p,q). Default to Cauchy-Schwarz unless the problem specifically involves other Lp norms.
Mean and Averaging Inequalities
These inequalities relate different types of averages and reveal how convexity shapes the behavior of functions applied to means.
Arithmetic Mean-Geometric Mean (AM-GM) Inequality
- For non-negative reals: nx1+x2+⋯+xn≥(x1x2⋯xn)1/n—arithmetic mean beats geometric mean
- Equality holds when all values are equal—this is your key to optimization problems
- Go-to tool for optimization: when maximizing a product subject to a sum constraint (or vice versa), AM-GM often gives the answer directly
Jensen's Inequality
- For convex functions f: f(nx1+⋯+xn)≤nf(x1)+⋯+f(xn)
- Inequality reverses for concave functions—always check the curvature before applying
- Explains why AM-GM works: since −ln(x) is convex, Jensen's applied to logarithms yields AM-GM
Compare: AM-GM vs. Jensen's—AM-GM is actually a special case of Jensen's (applied to the convex function −lnx). Use AM-GM for quick product/sum problems; use Jensen's when you're working with a specific convex or concave function and need to relate f(average) to average of f.
Growth and Approximation Inequalities
This inequality captures how exponential-type growth behaves, particularly useful for approximations and induction proofs.
Bernoulli's Inequality
- For x≥−1 and integer n≥0: (1+x)n≥1+nx
- Provides a linear lower bound on exponential growth—the bound is tight when x is small
- Perfect for induction proofs and establishing limits; often appears when bounding (1+n1)n type expressions
Probabilistic Bounding Inequalities
These inequalities let you control probabilities without knowing the full distribution—essential for worst-case analysis in statistics and probability.
Markov's Inequality
- For non-negative random variable X and a>0: P(X≥a)≤aE[X]
- Requires only the expected value—no variance or distribution shape needed
- Often gives weak bounds but works universally; it's your baseline tool when you know almost nothing about X
Chebyshev's Inequality
- Bounds deviation from the mean: P(∣X−μ∣≥kσ)≤k21
- Uses variance information to give tighter bounds than Markov—the tradeoff is you need to know σ2
- Distribution-free guarantee: no matter the shape, at least 1−k21 of probability lies within k standard deviations
Compare: Markov's vs. Chebyshev's—Markov requires only E[X] and works for any non-negative variable; Chebyshev requires E[X] and Var(X) but gives tighter bounds on deviation from the mean. If a problem gives you variance, use Chebyshev; if you only have the expected value, Markov is your only option.
Quick Reference Table
|
| Distance/metric properties | Triangle Inequality, Minkowski's Inequality |
| Bounding products by sums | Cauchy-Schwarz, Hölder's, Young's Inequality |
| Comparing means | AM-GM Inequality, Jensen's Inequality |
| Convexity applications | Jensen's Inequality |
| Exponential growth bounds | Bernoulli's Inequality |
| Probability tail bounds | Markov's Inequality, Chebyshev's Inequality |
| Lp space structure | Hölder's, Minkowski's Inequality |
| Optimization | AM-GM Inequality, Cauchy-Schwarz |
Self-Check Questions
-
Both Cauchy-Schwarz and Hölder's inequality bound sums of products. What condition on the exponents makes Cauchy-Schwarz a special case of Hölder's?
-
You're given a non-negative random variable with known mean but unknown variance. Which inequality can you use to bound P(X≥a), and what would you need to know to get a tighter bound?
-
Compare and contrast AM-GM and Jensen's inequality: How is AM-GM derived from Jensen's, and when would you prefer one over the other?
-
The triangle inequality and Minkowski's inequality both establish that "the norm of a sum is at most the sum of norms." What distinguishes their domains of application?
-
If an exam problem asks you to prove that a product ab is bounded above by a sum of powers of a and b, which inequality should you reach for first, and what condition must the exponents satisfy?