Properties of Probability
Probability properties give you the rules for how probabilities behave. They tell you what values probabilities can take, how they relate to each other, and how to combine them. Everything else in this course builds on these properties, so understanding them well now pays off later.
Fundamental Properties
Probability measures the likelihood of an event occurring, expressed as a number between 0 and 1.
There are a few core properties you need to know:
- Non-negativity: The probability of any event is greater than or equal to zero. You can't have a negative chance of something happening.
- Boundedness: Probabilities can't exceed 1. Combined with non-negativity, this means every probability lives in the range [0, 1].
- Normalization: The probabilities of all possible outcomes in the sample space must add up to 1. Something has to happen.
- Probability of the empty set: . An impossible event has zero probability.
- Probability of the sample space: . The certain event has probability 1.
Complement rule: For any event , the probability of not happening is . This works because either happens or it doesn't, and those two probabilities must sum to 1.
Example: The probability of rolling an even number on a fair die is . So the probability of rolling an odd number is .
Monotonicity: If event is a subset of event (meaning every outcome in is also in ), then . A smaller collection of outcomes can't be more likely than a larger one that contains it.
Example: Rolling a 6 on a fair die () is a subset of rolling an even number (). Sure enough, .
Kolmogorov's Axioms and Proofs
All probability properties can be derived from just three axioms, laid out by Kolmogorov:
- Axiom 1 (Non-negativity): for any event
- Axiom 2 (Normalization): , where is the sample space
- Axiom 3 (Additivity): For mutually exclusive events and ,
These three axioms are the foundation. Every other property is a consequence of them.

Proof of the Complement Rule
- and are mutually exclusive and together cover the entire sample space, so .
- By Axiom 3: .
- By Axiom 2: .
- Rearranging: .
Proof of the Inclusion-Exclusion Principle
For any two events and :
-
You can split into two mutually exclusive parts: and (the part of not in ).
-
By Axiom 3: .
-
Similarly, can be split into and , so .
-
Rearranging step 3: .
-
Substituting into step 2: .
The subtraction of corrects for double-counting the overlap between and .

Proof of Monotonicity
If , then .
- Write , where and are mutually exclusive.
- By Axiom 3: .
- By Axiom 1: .
- Therefore: .
Law of Total Probability
If events form a partition of the sample space (they're mutually exclusive and together cover all of ), then for any event :
This is useful when you can't calculate directly but you can figure it out by breaking it into cases.
Bayes' Theorem
The proof follows from the definition of conditional probability:
- By definition: .
- Also by definition: .
- Setting these equal and solving for gives Bayes' theorem.
Bayes' theorem lets you "reverse" a conditional probability. If you know , you can find .
Applying Probability Properties
These properties aren't just abstract rules. Here's how they show up in practice:
Complement rule in medical testing: If the probability of having a certain condition is , then the probability of not having it is . This is often easier than computing the "not having it" probability from scratch.
Normalization in weather forecasting: A forecast might assign probabilities of 60% sunny, 30% cloudy, and 10% rainy. These must sum to 100%. If they didn't, the model would be broken.
Additivity for system reliability: If a system has three independent failure modes with probabilities 0.01, 0.02, and 0.005, and these failures are mutually exclusive, the total failure probability is . (Note: for non-mutually-exclusive failures, you'd need inclusion-exclusion.)
Monotonicity in quality control: If an automated assembly line has a defect rate of 1.5% and a manual line has a defect rate of 3%, and the set of conditions causing defects in the automated line is a subset of those in the manual line, monotonicity confirms the automated line's probability should be lower.
Boundedness as a sanity check: Whenever you calculate a probability, check that your answer falls between 0 and 1. If you get or , something went wrong. This is one of the quickest ways to catch errors.