Fiveable

📊Honors Statistics Unit 5 Review

QR code for Honors Statistics practice questions

5.1 Continuous Probability Functions

5.1 Continuous Probability Functions

Written by the Fiveable Content Team • Last updated August 2025
Written by the Fiveable Content Team • Last updated August 2025
📊Honors Statistics
Unit & Topic Study Guides
Pep mascot

Continuous Probability Functions

Continuous probability functions model situations where a variable can take on infinitely many values within a range. Unlike discrete random variables (which you can count), continuous random variables like height, time, or temperature require a different approach: instead of assigning probability to individual outcomes, you calculate probability as the area under a curve.

Pep mascot
more resources to help you study

Continuous Probability Functions

Integration for continuous probabilities, Continuous Probability Functions | Introduction to Statistics – Gravina

Integration for continuous probabilities

A probability density function (PDF), denoted f(x)f(x), describes how probability is distributed across the values of a continuous random variable XX. Variables like time, weight, and temperature are common examples.

The core idea: to find the probability that XX falls within some range, you integrate the PDF over that range.

P(aXb)=abf(x)dxP(a \leq X \leq b) = \int_a^b f(x) \, dx

For example, if f(x)f(x) models the weight of a product, then 100150f(x)dx\int_{100}^{150} f(x) \, dx gives the probability that a randomly selected product weighs between 100 and 150 grams.

Every valid PDF must satisfy two properties:

  • Non-negativity: f(x)0f(x) \geq 0 for all xx. A density function can never be negative.
  • Total area equals 1: f(x)dx=1\int_{-\infty}^{\infty} f(x) \, dx = 1. This guarantees that the probabilities across all possible outcomes sum to certainty.

One more connection worth knowing: integrating the PDF from -\infty up to some value xx gives the cumulative distribution function (CDF), F(x)=P(Xx)F(x) = P(X \leq x). The CDF tells you the probability that the variable is at most xx.

Integration for continuous probabilities, Riemann integral - Wikipedia

Probability-area relationship in distributions

The link between probability and area is the single most important idea in this section. The area under the PDF curve between two x-values is the probability that the variable falls in that range. For instance, the area under a weight distribution curve between 60 and 70 kg equals the probability that a randomly chosen individual weighs between 60 and 70 kg.

Because the total area under the curve is 1, the probability of XX taking any single exact value is 0. This might feel strange, but think of it this way: there are infinitely many possible values in any interval, so the area above a single point (which has no width) is zero. A practical consequence is that P(X=5)P(X = 5) = 0, and P(aXb)=P(a<X<b)P(a \leq X \leq b) = P(a < X < b). Including or excluding endpoints doesn't change the probability.

The shape of the PDF curve tells you where values are more or less concentrated:

  • Higher values of f(x)f(x) mean the variable is more likely to fall near that region. The peak of a normal distribution, for example, sits at the mean.
  • Steeper slopes indicate probability density is changing rapidly, which you typically see in the tails of a distribution where values become increasingly rare.

Note that f(x)f(x) itself is not a probability. It's a density, and it can actually exceed 1. Only the area under the curve over an interval gives you a probability.

Applications of continuous probability functions

Solving a continuous probability problem generally follows these steps:

  1. Identify the distribution. Match the real-world scenario to a known distribution:

    • Normal (μ,σ)(\mu, \sigma): symmetric, bell-shaped data like heights or exam scores
    • Exponential (λ)(\lambda): time between independent events, like wait times at a bus stop or the lifespan of a light bulb
    • Continuous uniform (a,b)(a, b): all values equally likely between a minimum aa and maximum bb, like a random number generator on an interval
  2. Determine the parameters. Use the information given in the problem:

    • Normal: mean μ\mu and standard deviation σ\sigma
    • Exponential: rate parameter λ\lambda (the reciprocal of the mean)
    • Uniform: endpoints aa and bb
  3. Set up the integral. Translate the question into bounds of integration. For example, "probability a light bulb lasts between 1000 and 1500 hours" becomes 10001500f(x)dx\int_{1000}^{1500} f(x) \, dx using the appropriate PDF. For cumulative questions like "probability a student scores below the 60th percentile," you'd integrate from the lower bound of the distribution up to that score.

  4. Calculate and interpret. Compute the integral (by hand, table, or technology) and state what the result means in context. A result of 0.12 might mean there's a 12% chance a product is defective, or a 12% chance a customer waits more than 5 minutes.

Measures of central tendency and dispersion

Just as discrete distributions have expected values and variances, continuous distributions do too. The formulas use integration instead of summation.

Expected value (mean):

E(X)=xf(x)dxE(X) = \int_{-\infty}^{\infty} x \, f(x) \, dx

This is the long-run average value of XX if you could repeat the random process many times. It acts as the "balance point" of the distribution.

Variance:

Var(X)=(xμ)2f(x)dx\text{Var}(X) = \int_{-\infty}^{\infty} (x - \mu)^2 \, f(x) \, dx

Variance measures how spread out the distribution is around the mean. A larger variance means values are more dispersed.

Standard deviation is the square root of the variance, σ=Var(X)\sigma = \sqrt{\text{Var}(X)}, and it's in the same units as XX, which makes it easier to interpret than variance.

A useful computational shortcut for variance: Var(X)=E(X2)[E(X)]2\text{Var}(X) = E(X^2) - [E(X)]^2, where E(X2)=x2f(x)dxE(X^2) = \int_{-\infty}^{\infty} x^2 \, f(x) \, dx. This form is often faster than expanding (xμ)2(x - \mu)^2 inside the integral.