Study smarter with Fiveable
Get study guides, practice questions, and cheatsheets for all your subjects. Join 500,000+ students with a 96% pass rate.
Convex functions are the backbone of optimization theoryโand you'll encounter them constantly in problems involving minimization, duality, and constraint handling. The reason convexity matters so much is that any local minimum of a convex function is automatically a global minimum, which makes optimization tractable. When you're working through problems in convex geometry, you're being tested on your ability to recognize convexity, verify it using second derivatives or other criteria, and apply the right function to model a given scenario.
This guide organizes important convex functions by their structural properties and typical applications: elementary functions with simple curvature, functions built from set geometry, and specialized functions used in optimization algorithms. Don't just memorize definitionsโknow what makes each function convex and when you'd reach for it in a problem.
These are the foundational functions you'll use to build intuition about convexity. Their convexity follows directly from calculusโspecifically, non-negative second derivatives guarantee convexity for twice-differentiable functions.
Compare: Linear vs. Quadraticโboth have constant second derivatives, but linear functions have (affine) while quadratics have (strictly convex). If asked to find a unique minimizer, you need strict convexityโlinear functions won't give you that.
Some important convex functions arise by negating concave functions or applying transformations. Understanding this relationship helps you convert between convex and concave formulations in optimization problems.
Compare: Negative log vs. Entropyโboth derive convexity from the concavity of logarithms, but negative log operates on scalars while entropy operates on probability distributions. Entropy problems typically include the constraint .
Norms provide a natural way to measure size and distance, and their convexity follows from the triangle inequality and positive homogeneityโnot from second derivatives.
Compare: Norms vs. Support Functionsโboth are convex and positively homogeneous, but norms measure the "size" of a vector while support functions measure how far a set extends in a given direction. Support functions encode geometric information about sets.
These functions encode constraints geometrically, allowing you to incorporate feasibility conditions directly into the objective. They're essential for understanding constrained optimization through an unconstrained lens.
Compare: Indicator vs. Barrierโboth enforce feasibility, but indicator functions are discontinuous (hard constraints) while barrier functions are smooth (soft constraints that strengthen near boundaries). Interior-point methods use barriers; projected gradient methods work with indicators.
These functions maintain convexity under specific operations, allowing you to build complex convex functions from simpler ones while preserving optimization tractability.
Compare: Perspective vs. Original Functionโthe perspective transformation adds a scaling variable while maintaining convexity. This is useful when your problem naturally involves ratios or when you need to homogenize a function.
| Concept | Best Examples |
|---|---|
| Calculus-based convexity () | Quadratic, Exponential, Negative Log |
| Affine functions | Linear |
| Norm-based convexity | norm, norm, Support functions |
| Set-encoding functions | Indicator, Barrier |
| Probability/information theory | Entropy, Negative log |
| Convexity-preserving operations | Perspective functions |
| Regularization in ML | Norm functions (, ) |
Both linear and quadratic functions have constant second derivatives. Why does a quadratic function have a unique minimizer while a linear function (with ) does not?
Which two functions from this guide would you use to model a constrained optimization problem where you want iterates to stay strictly inside the feasible region? How do they differ in approach?
The logarithm is concave, yet entropy-based objectives appear in convex optimization. Explain how convexity is achieved in entropy minimization problems.
Compare norm functions and support functions: what geometric property do they share, and what different information do they encode about vectors versus sets?
You're given a convex function and need to create a joint function that remains convex while allowing the input to be scaled by . Which function type accomplishes this, and what condition must satisfy?