Probability density functions are the backbone of engineering analysisโthey're how you model uncertainty, predict system behavior, and make decisions when outcomes aren't deterministic. You're being tested on your ability to recognize which distribution fits which scenario, understand how parameters shape behavior, and apply the right PDF to problems involving reliability, quality control, hypothesis testing, and signal processing.
Don't just memorize the formulas. Know why each distribution exists, what real-world processes it models, and how changing parameters affects the shape. When you see an exam problem describing waiting times, failure rates, or sample statistics, you should immediately recognize which distribution family appliesโand understand the mathematical reasoning behind that choice.
Foundational Continuous Distributions
These distributions form the building blocks of probability theory. They model idealized scenarios and serve as the basis for more complex distributions.
Uniform Distribution
Equal probability across a bounded intervalโevery value between a and b is equally likely, with PDF f(x)=bโa1โ
Two parameters define the support: minimum a and maximum b, giving mean 2a+bโ and variance 12(bโa)2โ
Foundation for random number generationโtransforming uniform samples into other distributions is a core simulation technique
Normal (Gaussian) Distribution
Bell-shaped and symmetric around the mean ฮผโthe PDF is f(x)=ฯ2ฯโ1โeโ2ฯ2(xโฮผ)2โ
Defined by mean ฮผ and standard deviation ฯ, where approximately 68% of values fall within ยฑ1ฯ of the mean
Central Limit Theorem makes this universalโsums of independent random variables converge to normal, explaining why measurement errors and natural phenomena follow this distribution
Compare: Uniform vs. Normalโboth are symmetric, but uniform has bounded support with constant density while normal has unbounded support with density concentrated near the mean. If an FRQ asks about modeling "equally likely outcomes in a range," use uniform; for "accumulated random effects," use normal.
Time-to-Event and Reliability Distributions
These distributions model when something happensโfailure times, arrival processes, and system lifetimes. The key concept is the hazard rate (instantaneous failure probability).
Exponential Distribution
Models time until a single event with constant hazard rate ฮปโthe "memoryless" property means past time doesn't affect future probability
Single parameter ฮป (rate) gives mean ฮป1โ and PDF f(x)=ฮปeโฮปx for xโฅ0
Fundamental to queuing theory and reliabilityโuse this when failure rate doesn't change with age (electronic components, not mechanical wear)
Weibull Distribution
Generalizes exponential to handle varying failure ratesโshape parameter k determines whether hazard increases (k>1), decreases (k<1), or stays constant (k=1)
Two parameters: shape k and scale ฮป, with PDF f(x)=ฮปkโ(ฮปxโ)kโ1eโ(x/ฮป)k
Industry standard for reliability engineeringโmodels infant mortality (k<1), random failures (k=1), and wear-out (k>1) in a single framework
Gamma Distribution
Models waiting time for multiple eventsโif exponential is time to one event, gamma is time to k events
Two parameters: shape k (number of events) and scale ฮธ (or rate ฮฒ=1/ฮธ), with mean kฮธ
Reduces to exponential when k=1 and to chi-square when ฮธ=2 and k=ฮฝ/2โunderstanding these connections is frequently tested
Compare: Exponential vs. Weibullโexponential assumes constant failure rate (memoryless), while Weibull allows failure rate to change with time. On reliability problems, ask yourself: "Does age affect failure probability?" If yes, use Weibull.
Bounded and Proportion Distributions
When your random variable is constrained to a specific interval, these distributions apply. They're essential for modeling probabilities, percentages, and ratios.
Beta Distribution
Defined only on [0,1]โperfect for modeling probabilities, proportions, and Bayesian prior distributions
Two shape parameters ฮฑ and ฮฒ control asymmetry: ฮฑ>ฮฒ skews right, ฮฑ<ฮฒ skews left, ฮฑ=ฮฒ is symmetric
Extremely flexible: uniform (ฮฑ=ฮฒ=1), U-shaped (ฮฑ,ฮฒ<1), or unimodal (ฮฑ,ฮฒ>1)โthe conjugate prior for binomial likelihood in Bayesian inference
Lognormal Distribution
Models positive-only variables where multiplicative effects dominateโif ln(X) is normal, then X is lognormal
Parameters ฮผ and ฯ are the mean and standard deviation of ln(X), not of X itselfโa common exam trap
Right-skewed with heavy tailโmodels income distributions, stock prices, particle sizes, and any quantity that grows by percentages
Compare: Beta vs. Lognormalโboth can be right-skewed, but beta is bounded on [0,1] while lognormal is unbounded above. Use beta for proportions (market share, defect rates); use lognormal for positive quantities with multiplicative growth.
Sampling and Inference Distributions
These distributions arise from sampling processes and are essential for hypothesis testing, confidence intervals, and ANOVA. They're derived from normal distributions.
Chi-Square Distribution
Sum of squared standard normal variablesโif ZiโโผN(0,1), then โi=1kโZi2โโผฯk2โ
Single parameter: degrees of freedom k, with mean k and variance 2k; right-skewed for small k, approaches normal as kโโ
Primary use: variance testingโthe sample variance s2 follows a scaled chi-square, making this essential for confidence intervals on ฯ2
Student's t-Distribution
Ratio of normal to chi-squareโarises when estimating means with unknown population variance
Degrees of freedom ฮฝ control tail heaviness: smaller ฮฝ means heavier tails, ฮฝโโ approaches standard normal
Critical for small-sample inferenceโuse t-distribution instead of normal when n<30 and ฯ is unknown
F-Distribution
Ratio of two chi-square variablesโused to compare two variances or mean squares in ANOVA
Two parameters: d1โ (numerator df) and d2โ (denominator df)โorder matters, so Fd1โ,d2โโ๎ =Fd2โ,d1โโ
Right-skewed and positive-onlyโtest statistic for "is the variance ratio significantly different from 1?"
Compare: Chi-square vs. t vs. Fโall three derive from normal samples. Chi-square tests one variance, t tests one mean (unknown variance), F tests two variances or multiple means. Know which degrees of freedom formula applies to each test type.
Quick Reference Table
Concept
Best Examples
Bounded, equal probability
Uniform
Symmetric, sum of random effects
Normal
Time to single event (constant hazard)
Exponential
Time to multiple events
Gamma
Variable failure rates over time
Weibull
Proportions and probabilities on [0,1]
Beta
Positive-only, multiplicative growth
Lognormal
Variance testing
Chi-square
Mean testing (small samples)
Student's t
Comparing variances / ANOVA
F-distribution
Self-Check Questions
Which two distributions are memoryless, and what does this property mean mathematically?
You're modeling the proportion of defective items in a batch (values between 0 and 1). Which distribution is most appropriate, and what parameters would you adjust to reflect prior belief that defect rates are typically low?
Compare and contrast the chi-square and F-distributions: how are they mathematically related, and when would you use each in hypothesis testing?
A reliability engineer observes that component failure rates increase with age due to wear. Which distribution should they use, and what constraint on the shape parameter reflects this behavior?
If an FRQ gives you sample data and asks you to construct a confidence interval for the population mean with unknown variance, which distribution do you use for the critical valueโand how does your answer change as sample size grows large?