Probability mass functions (PMFs) are the backbone of discrete random variables. They assign probabilities to specific outcomes, helping us understand the likelihood of different events. PMFs are crucial for calculating probabilities, expected values, and variances in discrete scenarios.

In this part of the chapter, we'll learn how to define, construct, and use PMFs. We'll also explore their relationship with cumulative distribution functions (CDFs) and see how they apply to real-world problems involving discrete random variables.

Probability Mass Functions

Definition and Key Properties

Top images from around the web for Definition and Key Properties
Top images from around the web for Definition and Key Properties
  • (PMF) assigns probabilities to outcomes
  • Defined exclusively for discrete random variables with countable outcome sets
  • Sum of all probabilities in a PMF equals 1, representing total probability
  • PMFs always non-negative, assigning probabilities ≥ 0 to each outcome
  • of a PMF encompasses all possible values with non-zero probability
  • Adheres to probability axioms
    • Non-negativity: All probabilities ≥ 0
    • Additivity: Probability of union of disjoint events equals sum of individual probabilities
    • Normalization: Total probability equals 1

Mathematical Representation and Examples

  • Typically expressed as P(X=x)=p(x)P(X = x) = p(x) where X represents the random variable and x specific value
  • PMF example: p & \text{if } x = 1 \\ 1-p & \text{if } x = 0 \end{cases}$$
  • PMF example: p(x)=([n](https://www.fiveableKeyTerm:n)x)px(1p)nxp(x) = \binom{[n](https://www.fiveableKeyTerm:n)}{x} p^x (1-p)^{n-x} where n trials, x successes, p probability of success
  • PMF example: p(x)=eλλxx!p(x) = \frac{e^{-\lambda} \lambda^x}{x!} where average rate of events

Calculating Probabilities with PMFs

Single and Multiple Outcome Probabilities

  • Calculate single outcome probability by evaluating PMF at specific value
  • Compute probability for outcome ranges by summing PMF values over desired outcomes
  • Example: Fair six-sided die
    • P(X=3)=16P(X = 3) = \frac{1}{6} (single outcome)
    • P(X3)=P(X=1)+P(X=2)+P(X=3)=16+16+16=12P(X \leq 3) = P(X = 1) + P(X = 2) + P(X = 3) = \frac{1}{6} + \frac{1}{6} + \frac{1}{6} = \frac{1}{2} (multiple outcomes)
  • Use multiplication for independent events involving multiple random variables
  • Apply conditional probability formulas for dependent events

Expected Value and Variance Calculations

  • Compute (mean) using PMF E[X]=xxp(x)E[X] = \sum_{x} x \cdot p(x)
  • Calculate variance using PMF and expected value Var(X)=E[(XE[X])2]=x(xE[X])2p(x)Var(X) = E[(X - E[X])^2] = \sum_{x} (x - E[X])^2 \cdot p(x)
  • Derive standard deviation as square root of variance σ=Var(X)\sigma = \sqrt{Var(X)}
  • Example: Expected value for fair six-sided die E[X]=116+216+316+416+516+616=3.5E[X] = 1 \cdot \frac{1}{6} + 2 \cdot \frac{1}{6} + 3 \cdot \frac{1}{6} + 4 \cdot \frac{1}{6} + 5 \cdot \frac{1}{6} + 6 \cdot \frac{1}{6} = 3.5

Constructing PMFs

Methods for PMF Construction

  • Identify all possible outcomes of discrete random variable
  • Assign non-negative probabilities to each outcome, ensuring sum equals 1
  • Express PMF as function, often using piecewise notation for different variable ranges
  • Construct empirical PMF from data by calculating relative frequencies of observed outcomes
  • Example: Construct PMF for sum of two fair six-sided dice \frac{1}{36} & \text{if } x = 2 \text{ or } x = 12 \\ \frac{2}{36} & \text{if } x = 3 \text{ or } x = 11 \\ \frac{3}{36} & \text{if } x = 4 \text{ or } x = 10 \\ \frac{4}{36} & \text{if } x = 5 \text{ or } x = 9 \\ \frac{5}{36} & \text{if } x = 6 \text{ or } x = 8 \\ \frac{6}{36} & \text{if } x = 7 \end{cases}$$

Verification and Considerations

  • Verify constructed PMF satisfies all valid probability mass function properties
  • Ensure non-negativity of all probabilities
  • Confirm sum of all probabilities equals 1
  • Consider underlying distributions informing PMF shape (uniform, normal, exponential)
  • Examine symmetry or patterns in data to guide PMF construction
  • Example: Verify PMF for sum of two dice x=212p(x)=136+236+336+436+536+636+536+436+336+236+136=1\sum_{x=2}^{12} p(x) = \frac{1}{36} + \frac{2}{36} + \frac{3}{36} + \frac{4}{36} + \frac{5}{36} + \frac{6}{36} + \frac{5}{36} + \frac{4}{36} + \frac{3}{36} + \frac{2}{36} + \frac{1}{36} = 1

PMFs vs Cumulative Distribution Functions

Relationship and Conversion

  • Derive (CDF) from PMF by summing probabilities of values less than or equal to given value F(x)=P(Xx)=txp(t)F(x) = P(X \leq x) = \sum_{t \leq x} p(t)
  • CDF approaches 1 as random variable increases, non-decreasing function
  • Obtain PMF from CDF by taking difference between consecutive CDF values p(x)=F(x)F(x1)p(x) = F(x) - F(x-1)
  • Example: CDF for fair coin toss (H = 1, T = 0) 0 & \text{if } x < 0 \\ \frac{1}{2} & \text{if } 0 \leq x < 1 \\ 1 & \text{if } x \geq 1 \end{cases}$$

Characteristics and Applications

  • PMF provides probabilities for exact values, CDF for values less than or equal to given value
  • CDF continuous from right, including probability of current value
  • CDF for discrete random variables step function, jumps occur at PMF support values
  • Use CDF to calculate probabilities of ranges efficiently P(a<Xb)=F(b)F(a)P(a < X \leq b) = F(b) - F(a)
  • Example: Calculate probability of rolling 3 or less on fair die using CDF P(X3)=F(3)=16+16+16=12P(X \leq 3) = F(3) = \frac{1}{6} + \frac{1}{6} + \frac{1}{6} = \frac{1}{2}

Key Terms to Review (12)

Bernoulli Distribution: The Bernoulli distribution is a discrete probability distribution that models a single trial with two possible outcomes, typically labeled as success (1) and failure (0). It serves as the foundation for more complex distributions, such as the binomial distribution, which consists of multiple independent Bernoulli trials. Understanding this distribution is crucial for grasping various applications in statistics, especially in scenarios where outcomes can be modeled as yes/no or true/false.
Binomial Distribution: The binomial distribution models the number of successes in a fixed number of independent Bernoulli trials, each with the same probability of success. It is crucial for analyzing situations where there are two outcomes, like success or failure, and is directly connected to various concepts such as discrete random variables and probability mass functions.
Cumulative Distribution Function: The cumulative distribution function (CDF) of a random variable is a function that describes the probability that the variable will take a value less than or equal to a specific value. The CDF provides a complete description of the distribution of the random variable, allowing us to understand its behavior over time and its potential outcomes in both discrete and continuous contexts.
Discrete Random Variable: A discrete random variable is a type of variable that can take on a countable number of distinct values, often arising from counting processes. These variables are essential in probability because they allow us to model scenarios where outcomes are finite and measurable. Understanding discrete random variables is crucial for calculating probabilities, defining probability mass functions, and determining expected values and variances related to specific distributions.
Expected Value: Expected value is a fundamental concept in probability that represents the average outcome of a random variable, calculated as the sum of all possible values, each multiplied by their respective probabilities. It serves as a measure of the center of a probability distribution and provides insight into the long-term behavior of random variables, making it crucial for decision-making in uncertain situations.
Moment generating function: A moment generating function (MGF) is a mathematical tool used to characterize the probability distribution of a random variable by encapsulating all its moments. By taking the expected value of the exponential function of the random variable, the MGF provides a compact representation of the distribution and can be used to derive properties such as mean, variance, and higher moments. The MGF is particularly useful for working with both discrete and continuous distributions, and it relates closely to probability mass functions, probability generating functions, and various applications in statistical theory.
N: In probability and statistics, 'n' typically represents the number of trials or the sample size in an experiment or study. It is a crucial component in calculating probabilities, distributions, and the behavior of random variables, linking theoretical concepts to practical applications in statistical analysis.
Normalization Condition: The normalization condition is a fundamental principle in probability theory that ensures the total probability of all possible outcomes in a discrete probability distribution equals one. This requirement is crucial for probability mass functions, as it validates that the function adequately describes a legitimate probability distribution by confirming that all potential events have been accounted for.
Poisson Distribution: The Poisson distribution is a probability distribution that expresses the probability of a given number of events occurring in a fixed interval of time or space, given that these events occur with a known constant mean rate and are independent of the time since the last event. This distribution is particularly useful for modeling random events that happen at a constant average rate, which connects directly to the concept of discrete random variables and their characteristics.
Probability Mass Function: A probability mass function (PMF) is a function that gives the probability of each possible value of a discrete random variable. It assigns a probability to each outcome in the sample space, ensuring that the sum of all probabilities is equal to one. This concept is essential for understanding how probabilities are distributed among different values of a discrete random variable, which connects directly to the analysis of events, calculations of expected values, and properties of distributions.
Support: In probability, support refers to the set of values that a random variable can take on with non-zero probability. It identifies where the probability mass or density is concentrated, indicating the possible outcomes of a random variable. Understanding support is essential because it helps determine the range of values that contribute to the probability distribution, thereby influencing calculations and interpretations of probabilities.
λ: In the context of probability mass functions, λ (lambda) typically represents the average rate of occurrence or the expected value for a Poisson distribution. It quantifies how many events are expected to occur in a fixed interval of time or space, linking to various characteristics of discrete random variables and their distributions.
© 2024 Fiveable Inc. All rights reserved.
AP® and SAT® are trademarks registered by the College Board, which is not affiliated with, and does not endorse this website.