๐ŸƒEngineering Probability

Key Concepts of Probability Density Functions

Study smarter with Fiveable

Get study guides, practice questions, and cheatsheets for all your subjects. Join 500,000+ students with a 96% pass rate.

Get Started

Why This Matters

Probability density functions (PDFs) are how you model uncertainty and predict outcomes when results aren't deterministic. Mastering PDFs means being able to recognize which distribution fits which scenario, understand how parameters shape behavior, and apply the right PDF to problems involving reliability, quality control, hypothesis testing, and more.

Don't just memorize the formulas. Know why each distribution exists, what real-world processes it models, and how changing parameters affects the shape. When you see a problem describing waiting times, failure rates, or sample statistics, you should immediately recognize which distribution family applies.


Foundational Continuous Distributions

These distributions form the building blocks of probability theory. They model idealized scenarios and serve as the basis for more complex distributions.

Uniform Distribution

  • Equal probability across a bounded interval: every value between aa and bb is equally likely, with PDF f(x)=1bโˆ’af(x) = \frac{1}{b-a} for aโ‰คxโ‰คba \leq x \leq b
  • Two parameters define the support: minimum aa and maximum bb, giving mean a+b2\frac{a+b}{2} and variance (bโˆ’a)212\frac{(b-a)^2}{12}
  • Foundation for random number generation: transforming uniform samples into other distributions is a core simulation technique (this is called the inverse transform method)

Normal (Gaussian) Distribution

  • Bell-shaped and symmetric around the mean ฮผ\mu: the PDF is f(x)=1ฯƒ2ฯ€eโˆ’(xโˆ’ฮผ)22ฯƒ2f(x) = \frac{1}{\sigma\sqrt{2\pi}} e^{-\frac{(x-\mu)^2}{2\sigma^2}}
  • Defined by mean ฮผ\mu and standard deviation ฯƒ\sigma, where approximately 68% of values fall within ยฑ1ฯƒ\pm 1\sigma, 95% within ยฑ2ฯƒ\pm 2\sigma, and 99.7% within ยฑ3ฯƒ\pm 3\sigma of the mean
  • The Central Limit Theorem makes this universal: sums (and averages) of independent random variables converge to normal as the sample size grows, which explains why measurement errors and natural phenomena so often follow this distribution

Compare: Uniform vs. Normal: both are symmetric, but uniform has bounded support with constant density while normal has unbounded support with density concentrated near the mean. If a problem describes "equally likely outcomes in a range," use uniform; for "accumulated random effects," use normal.


Time-to-Event and Reliability Distributions

These distributions model when something happens: failure times, arrival processes, and system lifetimes. A key concept here is the hazard rate, which represents the instantaneous probability of an event occurring given that it hasn't occurred yet.

Exponential Distribution

  • Models time until a single event with constant hazard rate ฮป\lambda. The "memoryless" property means the probability of the event occurring in the next interval doesn't depend on how much time has already passed. Mathematically: P(X>s+tโˆฃX>s)=P(X>t)P(X > s + t \mid X > s) = P(X > t).
  • Single parameter ฮป\lambda (rate) gives mean 1ฮป\frac{1}{\lambda} and PDF f(x)=ฮปeโˆ’ฮปxf(x) = \lambda e^{-\lambda x} for xโ‰ฅ0x \geq 0
  • Fundamental to queuing theory and reliability: use this when the failure rate doesn't change with age (think electronic components, not mechanical wear)

Weibull Distribution

  • Generalizes exponential to handle varying failure rates: the shape parameter kk determines whether hazard increases (k>1k > 1), decreases (k<1k < 1), or stays constant (k=1k = 1, which reduces to exponential)
  • Two parameters: shape kk and scale ฮป\lambda, with PDF f(x)=kฮป(xฮป)kโˆ’1eโˆ’(x/ฮป)kf(x) = \frac{k}{\lambda}\left(\frac{x}{\lambda}\right)^{k-1} e^{-(x/\lambda)^k} for xโ‰ฅ0x \geq 0
  • Industry standard for reliability engineering: models infant mortality (k<1k < 1), random failures (k=1k = 1), and wear-out (k>1k > 1) in a single framework

Gamma Distribution

  • Models waiting time for multiple events: if exponential gives the time to one event, gamma gives the time to the kk-th event
  • Two parameters: shape kk and scale ฮธ\theta (or equivalently rate ฮฒ=1/ฮธ\beta = 1/\theta), with mean kฮธk\theta and variance kฮธ2k\theta^2
  • Connects to other distributions: it reduces to exponential when k=1k = 1, and to chi-square when ฮธ=2\theta = 2 and k=ฮฝ/2k = \nu/2. These connections come up often in problems.

Compare: Exponential vs. Weibull: exponential assumes a constant failure rate (memoryless), while Weibull allows the failure rate to change with time. On reliability problems, ask yourself: "Does age affect failure probability?" If yes, use Weibull.


Bounded and Proportion Distributions

When your random variable is constrained to a specific interval, these distributions apply. They're essential for modeling probabilities, percentages, and ratios.

Beta Distribution

  • Defined only on [0,1][0, 1]: perfect for modeling probabilities, proportions, and Bayesian prior distributions
  • Two shape parameters ฮฑ\alpha and ฮฒ\beta control asymmetry: ฮฑ>ฮฒ\alpha > \beta skews right (density concentrated toward 1), ฮฑ<ฮฒ\alpha < \beta skews left (density concentrated toward 0), and ฮฑ=ฮฒ\alpha = \beta is symmetric
  • Extremely flexible: it becomes uniform when ฮฑ=ฮฒ=1\alpha = \beta = 1, U-shaped when ฮฑ,ฮฒ<1\alpha, \beta < 1, or unimodal when ฮฑ,ฮฒ>1\alpha, \beta > 1. It's also the conjugate prior for binomial likelihood in Bayesian inference, which makes updating beliefs mathematically clean.

Lognormal Distribution

  • Models positive-only variables where multiplicative effects dominate: if lnโก(X)\ln(X) is normally distributed, then XX is lognormal
  • Parameters ฮผ\mu and ฯƒ\sigma are the mean and standard deviation of lnโก(X)\ln(X), not of XX itself. This is a common source of mistakes on problems. The actual mean of XX is eฮผ+ฯƒ2/2e^{\mu + \sigma^2/2}.
  • Right-skewed with a heavy tail: models income distributions, stock prices, particle sizes, and any quantity that grows by percentages rather than by fixed amounts

Compare: Beta vs. Lognormal: both can be right-skewed, but beta is bounded on [0,1][0, 1] while lognormal is unbounded above. Use beta for proportions (market share, defect rates); use lognormal for positive quantities with multiplicative growth.


Sampling and Inference Distributions

These distributions arise from sampling processes and are essential for hypothesis testing, confidence intervals, and ANOVA. They're all derived from normal distributions.

Chi-Square Distribution

  • Sum of squared standard normal variables: if ZiโˆผN(0,1)Z_i \sim N(0,1), then โˆ‘i=1kZi2โˆผฯ‡k2\sum_{i=1}^{k} Z_i^2 \sim \chi^2_k
  • Single parameter: degrees of freedom kk, with mean kk and variance 2k2k. The distribution is right-skewed for small kk and approaches normal as kโ†’โˆžk \to \infty.
  • Primary use is variance testing: the sample variance s2s^2 follows a scaled chi-square distribution, specifically (nโˆ’1)s2ฯƒ2โˆผฯ‡nโˆ’12\frac{(n-1)s^2}{\sigma^2} \sim \chi^2_{n-1}. This makes it essential for confidence intervals on ฯƒ2\sigma^2 and goodness-of-fit tests.

Student's t-Distribution

  • Ratio of a standard normal to the square root of a chi-square divided by its df: this arises naturally when you estimate a population mean but have to use the sample standard deviation instead of the true ฯƒ\sigma
  • Degrees of freedom ฮฝ\nu control tail heaviness: smaller ฮฝ\nu means heavier tails (more probability in the extremes), and as ฮฝโ†’โˆž\nu \to \infty the t-distribution approaches the standard normal
  • Critical for small-sample inference: use the t-distribution instead of the normal when ฯƒ\sigma is unknown and you're relying on ss. The common rule of thumb is that this matters most when n<30n < 30, though technically you should use t whenever ฯƒ\sigma is unknown.

F-Distribution

  • Ratio of two independent chi-square variables, each divided by its df: used to compare two variances or mean squares in ANOVA
  • Two parameters: d1d_1 (numerator df) and d2d_2 (denominator df): order matters, so Fd1,d2โ‰ Fd2,d1F_{d_1, d_2} \neq F_{d_2, d_1}
  • Right-skewed and positive-only: the test statistic answers "is the variance ratio significantly different from 1?"

Compare: Chi-square vs. t vs. F: all three derive from normal samples. Chi-square tests one variance, t tests one mean (with unknown variance), and F tests two variances or multiple means (via ANOVA). Know which degrees of freedom formula applies to each test type.


Quick Reference Table

ConceptBest Distribution
Bounded, equal probabilityUniform
Symmetric, sum of random effectsNormal
Time to single event (constant hazard)Exponential
Time to multiple eventsGamma
Variable failure rates over timeWeibull
Proportions and probabilities on [0,1][0,1]Beta
Positive-only, multiplicative growthLognormal
Variance testingChi-square
Mean testing (small samples, unknown ฯƒ\sigma)Student's t
Comparing variances / ANOVAF-distribution

Self-Check Questions

  1. Which two distributions are memoryless, and what does this property mean mathematically? (Hint: one is continuous, one is discrete.)

  2. You're modeling the proportion of defective items in a batch (values between 0 and 1). Which distribution is most appropriate, and what parameters would you adjust to reflect a prior belief that defect rates are typically low?

  3. Compare and contrast the chi-square and F-distributions: how are they mathematically related, and when would you use each in hypothesis testing?

  4. A reliability engineer observes that component failure rates increase with age due to wear. Which distribution should they use, and what constraint on the shape parameter reflects this behavior?

  5. You have sample data and need to construct a confidence interval for the population mean with unknown variance. Which distribution do you use for the critical value, and how does your answer change as sample size grows large?

Key Concepts of Probability Density Functions to Know for Intro to Probability