Fiveable

📊Honors Statistics Unit 9 Review

QR code for Honors Statistics practice questions

9.5 Additional Information and Full Hypothesis Test Examples

9.5 Additional Information and Full Hypothesis Test Examples

Written by the Fiveable Content Team • Last updated August 2025
Written by the Fiveable Content Team • Last updated August 2025
📊Honors Statistics
Unit & Topic Study Guides
Pep mascot

Hypothesis Testing Concepts

Hypothesis testing gives you a structured way to decide whether sample data supports or contradicts a claim about a population. Rather than guessing, you set up competing hypotheses, collect data, and use probability to determine how convincing the evidence is. This section pulls together the core concepts and walks through full examples from start to finish.

Pep mascot
more resources to help you study

Significance and p-value interpretation

The level of significance (α\alpha) is the threshold you set before running the test. It represents the maximum probability of a Type I error you're willing to accept, where a Type I error means rejecting H0H_0 when it's actually true. Common choices are 0.01, 0.05, and 0.10.

The p-value is the probability of obtaining a test statistic as extreme as (or more extreme than) what you observed, assuming H0H_0 is true. It comes from your sample data and gets compared directly to α\alpha:

  • Reject H0H_0 if p-value α\leq \alpha (the result is statistically significant)
  • Fail to reject H0H_0 if p-value >α> \alpha (the result is not statistically significant)

Think of α\alpha as the bar you set, and the p-value as the evidence trying to clear that bar. A smaller p-value means stronger evidence against H0H_0. For instance, a p-value of 0.001 is far more damning for H0H_0 than a p-value of 0.04, even though both lead to rejection at α=0.05\alpha = 0.05.

One more concept worth knowing: statistical power is the probability of correctly rejecting H0H_0 when it is actually false (i.e., detecting a real effect). Power increases with larger sample sizes and larger true effect sizes.

Types of hypothesis tests

The form of the alternative hypothesis (HaH_a) determines whether you're running a left-tailed, right-tailed, or two-tailed test. This matters because it dictates where the "rejection region" sits on the distribution.

  • Left-tailed: Ha:parameter<valueH_a: \text{parameter} < \text{value}. The critical region is in the left tail. Example: Ha:μ<100H_a: \mu < 100.
  • Right-tailed: Ha:parameter>valueH_a: \text{parameter} > \text{value}. The critical region is in the right tail. Example: Ha:p>0.5H_a: p > 0.5.
  • Two-tailed: Ha:parametervalueH_a: \text{parameter} \neq \text{value}. The critical region is split between both tails. Example: Ha:μ50H_a: \mu \neq 50.

How you find the p-value depends on the tail direction:

  • Left-tailed: p-value = area to the left of the test statistic
  • Right-tailed: p-value = area to the right of the test statistic
  • Two-tailed: p-value = twice the area in the more extreme tail

For the critical value approach (which you may also use), two-tailed tests split α\alpha across both tails, so each tail gets α/2\alpha/2. That gives you critical values of zα/2-z_{\alpha/2} and zα/2z_{\alpha/2} for a z-test.

Significance and p-value interpretation, A Closer Look at Tests of Significance | Boundless Statistics

Test Planning and Interpretation

  • A confidence interval complements a hypothesis test by giving a range of plausible values for the population parameter. If the hypothesized value falls outside the confidence interval, that's consistent with rejecting H0H_0.
  • Effect size quantifies the magnitude of the difference or relationship you're studying. A result can be statistically significant but have a tiny effect size, meaning it may not matter much in practice.
  • Sample size should be determined before collecting data. Larger samples increase power and precision, making it more likely you'll detect a real effect if one exists.

Full Hypothesis Test Examples

Significance and p-value interpretation, Introduction to Hypothesis Testing | Concepts in Statistics

Single population proportion test

Here's the general procedure, followed by a worked example.

The steps:

  1. State the hypotheses. Write H0:p=p0H_0: p = p_0 where p0p_0 is the claimed proportion. Then write HaH_a based on the research question: p<p0p < p_0, p>p0p > p_0, or pp0p \neq p_0.

  2. Set the significance level (α\alpha), typically given in the problem.

  3. Check conditions. Verify that np05np_0 \geq 5 and n(1p0)5n(1 - p_0) \geq 5 so the normal approximation is valid.

  4. Calculate the test statistic: z=p^p0p0(1p0)nz = \frac{\hat{p} - p_0}{\sqrt{\frac{p_0(1 - p_0)}{n}}}

where p^\hat{p} is the sample proportion and nn is the sample size. 5. Find the p-value using the standard normal distribution, based on the direction of HaH_a. 6. Make a decision. Reject H0H_0 if p-value α\leq \alpha; fail to reject if p-value >α> \alpha. 7. Interpret in context. State your conclusion in plain language, referring back to the original claim.

Worked example:

A school claims that 70% of its graduates attend a four-year college. A counselor suspects the true rate is lower. She surveys 150 recent graduates and finds that 96 attended a four-year college. Test at α=0.05\alpha = 0.05.

  1. H0:p=0.70H_0: p = 0.70 and Ha:p<0.70H_a: p < 0.70 (left-tailed, since the counselor suspects lower)

  2. α=0.05\alpha = 0.05

  3. Check conditions: np0=150(0.70)=1055np_0 = 150(0.70) = 105 \geq 5 and n(1p0)=150(0.30)=455n(1 - p_0) = 150(0.30) = 45 \geq 5. Conditions met.

  4. Sample proportion: p^=96150=0.64\hat{p} = \frac{96}{150} = 0.64

z=0.640.700.70(0.30)150=0.060.21150=0.060.0014=0.060.03741.60z = \frac{0.64 - 0.70}{\sqrt{\frac{0.70(0.30)}{150}}} = \frac{-0.06}{\sqrt{\frac{0.21}{150}}} = \frac{-0.06}{\sqrt{0.0014}} = \frac{-0.06}{0.0374} \approx -1.60

  1. Since this is left-tailed, the p-value is the area to the left of z=1.60z = -1.60. From the standard normal table, p-value 0.0548\approx 0.0548.
  2. Since 0.0548>0.050.0548 > 0.05, we fail to reject H0H_0.
  3. At the 0.05 significance level, there is not sufficient evidence to conclude that the proportion of graduates attending a four-year college is less than 70%.

Notice how close this was. The p-value barely missed the cutoff. That doesn't mean the claim is true; it means the sample didn't provide strong enough evidence to reject it at this threshold. A larger sample might yield a different result.

Single population mean test (z-test)

When the population standard deviation (σ\sigma) is known and the sample is large or the population is normal, you use a z-test for the mean.

The steps:

  1. State the hypotheses. H0:μ=μ0H_0: \mu = \mu_0 and Ha:μ<μ0H_a: \mu < \mu_0, μ>μ0\mu > \mu_0, or μμ0\mu \neq \mu_0.
  2. Set α\alpha.
  3. Calculate the test statistic: z=xˉμ0σ/nz = \frac{\bar{x} - \mu_0}{\sigma / \sqrt{n}}

where xˉ\bar{x} is the sample mean, μ0\mu_0 is the hypothesized mean, σ\sigma is the known population standard deviation, and nn is the sample size. 4. Find the p-value from the standard normal distribution. 5. Compare p-value to α\alpha and make a decision. 6. Interpret in context.

Worked example:

A battery manufacturer claims their batteries last an average of 500 hours. A consumer group tests 40 batteries and finds a sample mean of 490 hours. The population standard deviation is known to be 25 hours. Test at α=0.05\alpha = 0.05 whether the true mean is less than claimed.

  1. H0:μ=500H_0: \mu = 500 and Ha:μ<500H_a: \mu < 500 (left-tailed)
  2. α=0.05\alpha = 0.05
  3. Test statistic:

z=49050025/40=1025/6.3246=103.9532.53z = \frac{490 - 500}{25 / \sqrt{40}} = \frac{-10}{25 / 6.3246} = \frac{-10}{3.953} \approx -2.53

  1. Left-tailed p-value: area to the left of z=2.530.0057z = -2.53 \approx 0.0057.
  2. Since 0.0057<0.050.0057 < 0.05, we reject H0H_0.
  3. At the 0.05 significance level, there is sufficient evidence to conclude that the mean battery life is less than 500 hours. The consumer group's concern appears justified.

Common mistakes to avoid

  • Writing "accept H0H_0" instead of "fail to reject H0H_0." You never prove the null is true; you just didn't find enough evidence against it.
  • Using p^\hat{p} in the denominator of the proportion z-test formula. The standard error for hypothesis testing uses p0p_0 (the hypothesized value), not p^\hat{p}.
  • Forgetting to double the tail area for two-tailed tests. If your test statistic gives a one-tail area of 0.03, the two-tailed p-value is 0.06.
  • Confusing statistical significance with practical significance. A very large sample can make a tiny, meaningless difference statistically significant. Always consider effect size alongside your p-value.
  • Setting α\alpha after seeing the data. The significance level must be chosen before the test is conducted, or you risk biasing your conclusion.