Probability Distribution Functions (PDFs) for discrete random variables are key tools in statistics. They let you calculate the likelihood of specific outcomes in scenarios with countable possibilities, like rolling dice or counting defective items in a shipment.
Understanding PDFs is crucial for analyzing real-world data and making predictions. This section covers how to use these functions, how to distinguish between discrete and continuous variables, and how to verify whether a given function qualifies as a valid probability distribution.
Probability Distribution Function (PDF) for Discrete Random Variables

Fundamental Concepts
Before diving into PDFs, you need a few building blocks:
- Sample space: The set of all possible outcomes of a random experiment. For a single die roll, the sample space is {1, 2, 3, 4, 5, 6}.
- Probability axioms: The fundamental rules that all probabilities must follow. Every probability is between 0 and 1, and the probabilities of all outcomes in a sample space sum to 1.
- Independence: Two events are independent when the occurrence of one does not affect the probability of the other. Drawing two cards with replacement gives independent events; drawing without replacement does not.
- Conditional probability: The probability of an event occurring given that another event has already occurred, written .

Probability Calculations for Discrete Variables
A random variable is a function that assigns a numerical value to each outcome in the sample space. For example, if you roll two dice, you could define as the sum of the two faces.
The Probability Distribution Function (PDF) for a discrete random variable , denoted , gives the probability that takes on a specific value :
(You'll also see this called a probability mass function, or PMF. For discrete variables, the two terms refer to the same thing.)
To calculate probabilities using a PDF:
- Identify the random variable and list its possible values.
- Find the corresponding probability for each value using the given PDF.
- For the probability of multiple values, add the individual probabilities together. This works because discrete outcomes are mutually exclusive.
Example: Suppose has the following PDF:
| 1 | 2 | 3 | |
|---|---|---|---|
| 0.2 | 0.3 | 0.5 |

Discrete vs. Continuous Random Variables
This distinction matters because the type of variable determines how you find probabilities.
Discrete random variables take on a countable number of distinct values. They typically arise from counting: the number of defective items in a batch, the number of customers in a queue, or the number of heads in 10 coin flips. You find probabilities by summing the values of the PMF. A special case is the discrete uniform distribution, where every outcome has equal probability (like a fair die, where each face has probability ).
Continuous random variables take on an uncountable number of values within a range. They typically arise from measuring: the weight of a product, the time until a machine fails, or a person's height. Probabilities are found by integrating the density function over an interval, not by summing individual values. For a continuous variable, for any single point, which is a major conceptual difference from the discrete case.
Quick test: If you can list the possible values (even if the list is long), the variable is discrete. If the values fill an entire interval on the number line, it's continuous.
Validity of Discrete Probability Distributions
Not every function qualifies as a valid PDF. A function is a valid discrete probability distribution only if it satisfies two conditions:
- Non-negativity: for all values of . Probabilities can never be negative.
- Normalization: . The probabilities across all possible values must sum to exactly 1.
The function should also only assign nonzero probabilities to the actual possible values of the random variable (its support). Any value outside the support has .
To verify a distribution is valid:
- Check that every listed probability is non-negative.
- Add up all the probabilities and confirm the total equals 1.
- Confirm that no probability is assigned to values outside the variable's possible range.
Example: Is this a valid distribution?
| 0 | 1 | 2 | 3 | |
|---|---|---|---|---|
| 0.1 | 0.3 | 0.4 | 0.2 |
- All values are . ✓
- . ✓
This is a valid distribution. If the sum had been anything other than 1, or if any value had been negative, it would fail.