Fiveable

๐Ÿคน๐ŸผFormal Logic II Unit 11 Review

QR code for Formal Logic II practice questions

11.2 Probability theory and Bayesian inference

11.2 Probability theory and Bayesian inference

Written by the Fiveable Content Team โ€ข Last updated August 2025
Written by the Fiveable Content Team โ€ข Last updated August 2025
๐Ÿคน๐ŸผFormal Logic II
Unit & Topic Study Guides

Fundamentals of Probability Theory

Basic Concepts and Definitions

Probability measures the likelihood that an event will occur, expressed as a number between 0 and 1.

  • 0 indicates impossibility (e.g., rolling a 7 on a standard six-sided die)
  • 1 indicates certainty (e.g., drawing a card that is either red or black from a standard deck)

The probability of an event AA is denoted P(A)P(A). In the classical interpretation, you calculate it by dividing the number of favorable outcomes by the total number of equally likely outcomes:

P(A)=numberย ofย favorableย outcomestotalย numberย ofย possibleย outcomesP(A) = \frac{\text{number of favorable outcomes}}{\text{total number of possible outcomes}}

For example, rolling a fair six-sided die, the probability of getting a 3 is P(3)=16P(3) = \frac{1}{6}.

The complement of an event AA, written ยฌA\neg A (or sometimes Aโ€ฒA'), is the event that AA does not occur. Its probability is:

P(ยฌA)=1โˆ’P(A)P(\neg A) = 1 - P(A)

If the probability of a coin landing heads is 0.5, the probability of not getting heads is 1โˆ’0.5=0.51 - 0.5 = 0.5.

Rules for Combining Probabilities

Addition rule (disjunction). The probability of either AA or BB occurring is:

P(AโˆชB)=P(A)+P(B)โˆ’P(AโˆฉB)P(A \cup B) = P(A) + P(B) - P(A \cap B)

You subtract P(AโˆฉB)P(A \cap B) to avoid double-counting outcomes where both events happen. For example, the probability of drawing a heart or a king from a standard 52-card deck:

P(heart)+P(king)โˆ’P(heartย andย king)=1352+452โˆ’152=1652P(\text{heart}) + P(\text{king}) - P(\text{heart and king}) = \frac{13}{52} + \frac{4}{52} - \frac{1}{52} = \frac{16}{52}

The king of hearts gets counted in both "hearts" and "kings," so you subtract it once.

Multiplication rule (conjunction). The probability of both AA and BB occurring is:

P(AโˆฉB)=P(A)ร—P(BโˆฃA)P(A \cap B) = P(A) \times P(B|A)

Here P(BโˆฃA)P(B|A) is the conditional probability of BB given that AA has occurred. For example, the probability of drawing two hearts in a row from a standard deck (without replacement):

P(1stย heart)ร—P(2ndย heartโˆฃ1stย heart)=1352ร—1251=1562652P(\text{1st heart}) \times P(\text{2nd heart} | \text{1st heart}) = \frac{13}{52} \times \frac{12}{51} = \frac{156}{2652}

After drawing one heart, only 12 hearts remain among 51 cards.

Independence. Two events AA and BB are independent if the occurrence of one does not affect the probability of the other. When events are independent, the multiplication rule simplifies to:

P(AโˆฉB)=P(A)ร—P(B)P(A \cap B) = P(A) \times P(B)

Flipping a coin and rolling a die are independent events, so the probability of heads and a 6 is 0.5ร—16=1120.5 \times \frac{1}{6} = \frac{1}{12}.

Bayes' Theorem for Updating Probabilities

Bayes' Theorem Formula and Components

Bayes' theorem lets you reverse the direction of a conditional probability. If you know how likely the evidence is given a hypothesis, you can calculate how likely the hypothesis is given the evidence. The formula:

P(HโˆฃE)=P(EโˆฃH)ร—P(H)P(E)P(H|E) = \frac{P(E|H) \times P(H)}{P(E)}

Each component has a standard name:

  • P(H)P(H) โ€” the prior probability of hypothesis HH before seeing the evidence
  • P(EโˆฃH)P(E|H) โ€” the likelihood, i.e., how probable the evidence is if HH is true
  • P(E)P(E) โ€” the marginal probability of the evidence (across all hypotheses)
  • P(HโˆฃE)P(H|E) โ€” the posterior probability of HH after taking the evidence into account
Basic concepts and definitions, Bayesian Probability Illustration Diagram | TikZ example

Worked Example: Medical Diagnosis

Suppose a disease affects 1 in 1,000 people. A test for the disease has a 99% true positive rate (sensitivity) and a 2% false positive rate. You test positive. What's the probability you actually have the disease?

  1. Identify the prior. P(disease)=0.001P(\text{disease}) = 0.001, so P(ยฌdisease)=0.999P(\neg \text{disease}) = 0.999.
  2. Identify the likelihood. P(positiveโˆฃdisease)=0.99P(\text{positive} | \text{disease}) = 0.99.
  3. Calculate the marginal probability of a positive result. This accounts for all ways you could test positive (true positives + false positives):

P(positive)=P(positiveโˆฃdisease)ร—P(disease)+P(positiveโˆฃยฌdisease)ร—P(ยฌdisease)P(\text{positive}) = P(\text{positive} | \text{disease}) \times P(\text{disease}) + P(\text{positive} | \neg \text{disease}) \times P(\neg \text{disease})

=(0.99)(0.001)+(0.02)(0.999)=0.00099+0.01998=0.02097= (0.99)(0.001) + (0.02)(0.999) = 0.00099 + 0.01998 = 0.02097

  1. Apply Bayes' theorem:

P(diseaseโˆฃpositive)=0.99ร—0.0010.02097โ‰ˆ0.047P(\text{disease} | \text{positive}) = \frac{0.99 \times 0.001}{0.02097} \approx 0.047

The result is roughly 4.7%. Even with a positive test, there's only about a 1 in 21 chance you have the disease. This is counterintuitive, but it makes sense: the disease is so rare that false positives vastly outnumber true positives.

Updating Probabilities with New Evidence

The real power of Bayesian reasoning is iterative updating. Each posterior can become the prior for the next round of evidence.

  • The prior represents your initial belief before new data arrives. In the example above, it was the disease prevalence (0.001).
  • The likelihood quantifies how well each hypothesis predicts the observed evidence. A test with higher sensitivity produces a larger likelihood for the disease hypothesis.
  • The marginal probability normalizes everything so the posterior is a valid probability. You compute it by summing P(EโˆฃHi)ร—P(Hi)P(E|H_i) \times P(H_i) across all competing hypotheses.

If the person in the example above takes a second independent test and again tests positive, you'd plug the new prior of 0.047 into Bayes' theorem with the same likelihood and false positive rate. The posterior would jump significantly higher, because now the prior is no longer tiny.

Conditional Probabilities and Inductive Reasoning

Understanding Conditional Probabilities

Conditional probability is the probability of event AA occurring given that event BB has already occurred:

P(AโˆฃB)=P(AโˆฉB)P(B)P(A|B) = \frac{P(A \cap B)}{P(B)}

For example, the probability of drawing a red card given that a face card has been drawn from a standard deck: there are 12 face cards total, 6 of which are red, so P(redโˆฃfaceย card)=612=12P(\text{red} | \text{face card}) = \frac{6}{12} = \frac{1}{2}.

Conditional probabilities matter for inductive reasoning because they capture how evidence shifts the plausibility of a hypothesis. Observing BB changes the relevant sample space from "all outcomes" to "outcomes where BB is true," which can dramatically change the probability of AA.

This is exactly what Bayes' theorem formalizes: you start with P(H)P(H), observe evidence EE, and compute P(HโˆฃE)P(H|E). The conditional probability structure is what makes the update work.

Basic concepts and definitions, Why It Matters: Inference for One Proportion | Concepts in Statistics

Conditional Independence

Two events AA and BB are conditionally independent given a third event CC if:

P(AโˆฉBโˆฃC)=P(AโˆฃC)ร—P(BโˆฃC)P(A \cap B | C) = P(A|C) \times P(B|C)

This means that once you know CC, learning AA tells you nothing new about BB (and vice versa). Events can be conditionally independent given CC even if they are not independent overall.

For example, consider two different diagnostic tests for the same disease. The test results might be correlated in the general population (both are more likely to be positive for sick people). But given that you know whether the patient has the disease, the two test results might be independent of each other, since each test responds to the disease through a different mechanism. This conditional independence assumption is what justifies treating successive test results as independent updates in Bayesian reasoning.

Fallacies and Biases in Probabilistic Reasoning

Common Fallacies in Probability

Base rate neglect occurs when someone ignores the prior probability of an event and focuses only on the specific evidence. The medical test example above illustrates this perfectly: people intuitively feel that a 99%-accurate positive test means they almost certainly have the disease, but they're neglecting the base rate (only 1 in 1,000 people are affected). Bayes' theorem is the corrective here.

The conjunction fallacy is the mistaken belief that two events occurring together is more probable than either event alone. This violates a basic axiom: P(AโˆฉB)โ‰คP(A)P(A \cap B) \leq P(A). The classic example comes from Tversky and Kahneman's "Linda problem": participants judged it more likely that Linda is both a bank teller and a feminist activist than that she is a bank teller, because the conjunction fit a narrative better. Narratives aren't probabilities.

The gambler's fallacy is the belief that past outcomes influence future independent events. After flipping ten heads in a row with a fair coin, many people feel tails is "due." But each flip is independent: P(heads)=0.5P(\text{heads}) = 0.5 regardless of prior results. The coin has no memory.

Cognitive Biases Affecting Probability Judgments

Confirmation bias is the tendency to seek, interpret, and remember evidence that supports your existing beliefs while discounting evidence that contradicts them. In Bayesian terms, this is like selectively choosing which evidence to update on, or inflating the likelihood for your favored hypothesis. The antidote is to actively consider: how probable is this evidence if my hypothesis is wrong?

The availability heuristic leads people to estimate probabilities based on how easily examples come to mind. Plane crashes receive heavy media coverage, so people overestimate their frequency relative to, say, car accidents. The ease of recall is not a reliable guide to actual frequency.

Both fallacies and biases point to the same lesson: human intuition about probability is unreliable. The formal machinery of probability theory and Bayes' theorem exists precisely to discipline that intuition. Recognizing where your reasoning departs from the math is a core skill in inductive logic.