Fiveable

🥖Linear Modeling Theory Unit 3 Review

QR code for Linear Modeling Theory practice questions

3.1 Hypothesis Testing for Regression Coefficients

3.1 Hypothesis Testing for Regression Coefficients

Written by the Fiveable Content Team • Last updated August 2025
Written by the Fiveable Content Team • Last updated August 2025
🥖Linear Modeling Theory
Unit & Topic Study Guides

Hypothesis Testing for Regression Coefficients

Hypothesis testing for regression coefficients lets you determine whether a predictor variable has a statistically significant linear relationship with the response variable, or whether the observed association could plausibly be due to chance. This is the foundation of inference in simple linear regression: without it, you have estimates but no way to judge whether those estimates reflect real patterns in the population.

Formulating Null and Alternative Hypotheses

The null hypothesis claims that the true population coefficient is zero, meaning the predictor has no linear effect on the response:

H0:βi=0H_0: \beta_i = 0

where βi\beta_i is the population coefficient for the ithi^{th} predictor. If this is true, knowing the predictor's value tells you nothing (linearly) about the response.

The alternative hypothesis claims the coefficient is not zero, meaning a linear relationship does exist:

Ha:βi0H_a: \beta_i \neq 0

This is a two-sided test. Sometimes you have a directional prediction, which calls for a one-sided alternative:

  • Ha:βi>0H_a: \beta_i > 0 if you expect a positive relationship
  • Ha:βi<0H_a: \beta_i < 0 if you expect a negative relationship

For example, if a researcher believes that increased advertising expenditure leads to higher sales, the appropriate alternative would be Ha:βadvertising>0H_a: \beta_{\text{advertising}} > 0. One-sided tests are more powerful in the hypothesized direction but can't detect effects in the opposite direction, so only use them when you have a strong prior reason.

Conducting Hypothesis Tests Using t-Tests

The t-test for a regression coefficient compares the estimated coefficient to the variability you'd expect from sampling alone. Here's the process:

  1. Compute the test statistic. The t-statistic measures how many standard errors the estimate is from zero:

t=β^i0SE(β^i)=β^iSE(β^i)t = \frac{\hat{\beta}_i - 0}{SE(\hat{\beta}_i)} = \frac{\hat{\beta}_i}{SE(\hat{\beta}_i)}

where β^i\hat{\beta}_i is the estimated coefficient and SE(β^i)SE(\hat{\beta}_i) is its standard error.

  1. Understand the standard error. The standard error of β^i\hat{\beta}_i captures how much the estimate would vary across repeated samples. It depends on the variance of the residuals and the spread of the predictor values. Larger residual variance inflates the standard error; greater spread in the predictor shrinks it.

  2. Determine the degrees of freedom. For the t-distribution, degrees of freedom equal np1n - p - 1, where nn is the number of observations and pp is the number of predictors. In simple linear regression, p=1p = 1, so degrees of freedom are n2n - 2.

  3. Compare to the critical value or compute a p-value. If the absolute value of the t-statistic exceeds the critical value for your chosen α\alpha and degrees of freedom, you reject H0H_0. For instance, with α=0.05\alpha = 0.05, n=50n = 50, and p=1p = 1, the degrees of freedom are 5011=4850 - 1 - 1 = 48, and the two-tailed critical value is approximately ±2.011\pm 2.011.

Note: Since this is a simple linear regression course, you'll typically have p=1p = 1. The formula np1n - p - 1 generalizes to multiple regression, but for now, think of it as n2n - 2.

Formulating Null and Alternative Hypotheses, Setting up Hypothesis Tests

Interpreting Hypothesis Test Results

Rejecting or Failing to Reject the Null Hypothesis

If you reject H0H_0: There is sufficient evidence that βi0\beta_i \neq 0. The predictor has a statistically significant linear relationship with the response. Changes in the predictor are associated with changes in the response beyond what you'd expect from random variation alone.

If you fail to reject H0H_0: The data don't provide strong enough evidence to conclude that βi0\beta_i \neq 0. This does not mean the predictor is unrelated to the response. It means the signal wasn't strong enough relative to the noise in your data. A small sample size, high variability, or a genuinely weak effect could all produce this result.

Formulating Null and Alternative Hypotheses, Chapter 8. Regression Basics – Introductory Business Statistics with Interactive Spreadsheets ...

Understanding the Coefficient's Sign and Magnitude

The sign of β^i\hat{\beta}_i tells you the direction of the linear relationship:

  • Positive coefficient: as the predictor increases, the response tends to increase (direct relationship)
  • Negative coefficient: as the predictor increases, the response tends to decrease (inverse relationship)

The magnitude tells you the size of the effect. Specifically, β^i\hat{\beta}_i represents the expected change in the response for a one-unit increase in the predictor, holding all else constant. If the coefficient for "age" is 0.5, then each additional year of age is associated with a 0.5-unit increase in the response.

Keep in mind that a coefficient can be large in magnitude but not statistically significant (if the standard error is also large), or small in magnitude but highly significant (if estimated very precisely). Statistical significance and practical importance are separate questions.

Significance of Regression Coefficients

Using P-Values to Determine Significance

The p-value is the probability of observing a t-statistic as extreme as (or more extreme than) the one you calculated, assuming the null hypothesis is true. It quantifies how surprising your data are under H0H_0.

  • A small p-value (less than α\alpha) means the observed result would be unlikely if βi\beta_i were truly zero. You reject H0H_0 and conclude the coefficient is statistically significant. For example, if α=0.05\alpha = 0.05 and the p-value is 0.02, you reject H0H_0.
  • A large p-value (greater than α\alpha) means the observed result is reasonably consistent with βi=0\beta_i = 0. You fail to reject H0H_0. For example, if α=0.05\alpha = 0.05 and the p-value is 0.15, you do not reject H0H_0.

The p-value does not tell you the probability that H0H_0 is true. It tells you how likely the observed data (or something more extreme) would be if H0H_0 were true. That distinction matters.

Choosing an Appropriate Significance Level

The significance level α\alpha is the threshold you set before testing. It represents the maximum probability of a Type I error, which is rejecting H0H_0 when it's actually true (a false positive).

Common choices:

  • α=0.01\alpha = 0.01: Stringent. Requires very strong evidence to reject H0H_0. Used when false positives carry serious consequences, such as in medical research or regulatory decisions.
  • α=0.05\alpha = 0.05: The most common default across many fields. Balances the risk of false positives and false negatives.
  • α=0.10\alpha = 0.10: More lenient. Useful in exploratory analyses or when missing a real effect (Type II error) is more costly than a false alarm.

There's a tradeoff: lowering α\alpha reduces your Type I error rate but increases your Type II error rate (failing to detect a real effect). The right choice depends on the stakes involved and the goals of your analysis.