Probability density functions are the backbone of engineering analysis—they're how you model uncertainty, predict system behavior, and make decisions when outcomes aren't deterministic. You're being tested on your ability to recognize which distribution fits which scenario, understand how parameters shape behavior, and apply the right PDF to problems involving reliability, quality control, hypothesis testing, and signal processing.
Don't just memorize the formulas. Know why each distribution exists, what real-world processes it models, and how changing parameters affects the shape. When you see an exam problem describing waiting times, failure rates, or sample statistics, you should immediately recognize which distribution family applies—and understand the mathematical reasoning behind that choice.
Foundational Continuous Distributions
These distributions form the building blocks of probability theory. They model idealized scenarios and serve as the basis for more complex distributions.
Uniform Distribution
Equal probability across a bounded interval—every value between a and b is equally likely, with PDF f(x)=b−a1
Two parameters define the support: minimum a and maximum b, giving mean 2a+b and variance 12(b−a)2
Foundation for random number generation—transforming uniform samples into other distributions is a core simulation technique
Normal (Gaussian) Distribution
Bell-shaped and symmetric around the mean μ—the PDF is f(x)=σ2π1e−2σ2(x−μ)2
Defined by mean μ and standard deviation σ, where approximately 68% of values fall within ±1σ of the mean
Central Limit Theorem makes this universal—sums of independent random variables converge to normal, explaining why measurement errors and natural phenomena follow this distribution
Compare: Uniform vs. Normal—both are symmetric, but uniform has bounded support with constant density while normal has unbounded support with density concentrated near the mean. If an FRQ asks about modeling "equally likely outcomes in a range," use uniform; for "accumulated random effects," use normal.
Time-to-Event and Reliability Distributions
These distributions model when something happens—failure times, arrival processes, and system lifetimes. The key concept is the hazard rate (instantaneous failure probability).
Exponential Distribution
Models time until a single event with constant hazard rate λ—the "memoryless" property means past time doesn't affect future probability
Single parameter λ (rate) gives mean λ1 and PDF f(x)=λe−λx for x≥0
Fundamental to queuing theory and reliability—use this when failure rate doesn't change with age (electronic components, not mechanical wear)
Weibull Distribution
Generalizes exponential to handle varying failure rates—shape parameter k determines whether hazard increases (k>1), decreases (k<1), or stays constant (k=1)
Two parameters: shape k and scale λ, with PDF f(x)=λk(λx)k−1e−(x/λ)k
Industry standard for reliability engineering—models infant mortality (k<1), random failures (k=1), and wear-out (k>1) in a single framework
Gamma Distribution
Models waiting time for multiple events—if exponential is time to one event, gamma is time to k events
Two parameters: shape k (number of events) and scale θ (or rate β=1/θ), with mean kθ
Reduces to exponential when k=1 and to chi-square when θ=2 and k=ν/2—understanding these connections is frequently tested
Compare: Exponential vs. Weibull—exponential assumes constant failure rate (memoryless), while Weibull allows failure rate to change with time. On reliability problems, ask yourself: "Does age affect failure probability?" If yes, use Weibull.
Bounded and Proportion Distributions
When your random variable is constrained to a specific interval, these distributions apply. They're essential for modeling probabilities, percentages, and ratios.
Beta Distribution
Defined only on [0,1]—perfect for modeling probabilities, proportions, and Bayesian prior distributions
Two shape parameters α and β control asymmetry: α>β skews right, α<β skews left, α=β is symmetric
Extremely flexible: uniform (α=β=1), U-shaped (α,β<1), or unimodal (α,β>1)—the conjugate prior for binomial likelihood in Bayesian inference
Lognormal Distribution
Models positive-only variables where multiplicative effects dominate—if ln(X) is normal, then X is lognormal
Parameters μ and σ are the mean and standard deviation of ln(X), not of X itself—a common exam trap
Right-skewed with heavy tail—models income distributions, stock prices, particle sizes, and any quantity that grows by percentages
Compare: Beta vs. Lognormal—both can be right-skewed, but beta is bounded on [0,1] while lognormal is unbounded above. Use beta for proportions (market share, defect rates); use lognormal for positive quantities with multiplicative growth.
Sampling and Inference Distributions
These distributions arise from sampling processes and are essential for hypothesis testing, confidence intervals, and ANOVA. They're derived from normal distributions.
Chi-Square Distribution
Sum of squared standard normal variables—if Zi∼N(0,1), then ∑i=1kZi2∼χk2
Single parameter: degrees of freedom k, with mean k and variance 2k; right-skewed for small k, approaches normal as k→∞
Primary use: variance testing—the sample variance s2 follows a scaled chi-square, making this essential for confidence intervals on σ2
Student's t-Distribution
Ratio of normal to chi-square—arises when estimating means with unknown population variance
Degrees of freedom ν control tail heaviness: smaller ν means heavier tails, ν→∞ approaches standard normal
Critical for small-sample inference—use t-distribution instead of normal when n<30 and σ is unknown
F-Distribution
Ratio of two chi-square variables—used to compare two variances or mean squares in ANOVA
Two parameters: d1 (numerator df) and d2 (denominator df)—order matters, so Fd1,d2=Fd2,d1
Right-skewed and positive-only—test statistic for "is the variance ratio significantly different from 1?"
Compare: Chi-square vs. t vs. F—all three derive from normal samples. Chi-square tests one variance, t tests one mean (unknown variance), F tests two variances or multiple means. Know which degrees of freedom formula applies to each test type.
Quick Reference Table
Concept
Best Examples
Bounded, equal probability
Uniform
Symmetric, sum of random effects
Normal
Time to single event (constant hazard)
Exponential
Time to multiple events
Gamma
Variable failure rates over time
Weibull
Proportions and probabilities on [0,1]
Beta
Positive-only, multiplicative growth
Lognormal
Variance testing
Chi-square
Mean testing (small samples)
Student's t
Comparing variances / ANOVA
F-distribution
Self-Check Questions
Which two distributions are memoryless, and what does this property mean mathematically?
You're modeling the proportion of defective items in a batch (values between 0 and 1). Which distribution is most appropriate, and what parameters would you adjust to reflect prior belief that defect rates are typically low?
Compare and contrast the chi-square and F-distributions: how are they mathematically related, and when would you use each in hypothesis testing?
A reliability engineer observes that component failure rates increase with age due to wear. Which distribution should they use, and what constraint on the shape parameter reflects this behavior?
If an FRQ gives you sample data and asks you to construct a confidence interval for the population mean with unknown variance, which distribution do you use for the critical value—and how does your answer change as sample size grows large?