Fiveable

📈Intro to Probability for Business Unit 11 Review

QR code for Intro to Probability for Business practice questions

11.3 Least Squares Method and Regression Equation

11.3 Least Squares Method and Regression Equation

Written by the Fiveable Content Team • Last updated August 2025
Written by the Fiveable Content Team • Last updated August 2025
📈Intro to Probability for Business
Unit & Topic Study Guides

Simple linear regression is a powerful tool for understanding relationships between variables. It helps us predict one variable based on another, like guessing someone's blood pressure from their age. The method uses a straight line to represent this relationship.

The least squares method finds the best-fitting line by minimizing differences between observed and predicted values. We can use this line to make predictions, but it's important to remember that these predictions aren't perfect and come with some uncertainty.

Simple Linear Regression

Concept of simple linear regression

  • Statistical method models relationship between two variables
    • Independent (explanatory) variable: xx (age, income)
    • Dependent (response) variable: yy (blood pressure, spending)
  • Determines strength and direction of linear relationship between variables
  • Predicts value of dependent variable based on independent variable
  • Relationship represented by straight line called regression line
    • Equation: y=β0+β1x+ϵy = \beta_0 + \beta_1x + \epsilon
      • β0\beta_0: y-intercept, value of yy when x=0x = 0
      • β1\beta_1: slope, change in yy for one-unit change in xx
      • ϵ\epsilon: random error term, deviation of observed values from regression line
Concept of simple linear regression, Simple linear regression - Wikipedia

Least squares method application

  • Estimates parameters β0\beta_0 and β1\beta_1 of regression line
    • Minimizes sum of squared differences between observed and predicted yy values
  • Estimate slope β1\beta_1 using formula: β1^=i=1n(xixˉ)(yiyˉ)i=1n(xixˉ)2\hat{\beta_1} = \frac{\sum_{i=1}^{n} (x_i - \bar{x})(y_i - \bar{y})}{\sum_{i=1}^{n} (x_i - \bar{x})^2}
    • β1^\hat{\beta_1}: estimated slope
    • xix_i, yiy_i: observed values of independent and dependent variables
    • xˉ\bar{x}, yˉ\bar{y}: sample means of independent and dependent variables
    • nn: number of observations
  • Estimate y-intercept β0\beta_0 using formula: β0^=yˉβ1^xˉ\hat{\beta_0} = \bar{y} - \hat{\beta_1}\bar{x}
    • β0^\hat{\beta_0}: estimated y-intercept
Concept of simple linear regression, Linear Regression (2 of 4) | Concepts in Statistics

Derivation of regression equation

  • Substitute estimated values of β0\beta_0 and β1\beta_1 into general equation: y=β0+β1x+ϵy = \beta_0 + \beta_1x + \epsilon
  • Resulting regression equation: y^=β0^+β1^x\hat{y} = \hat{\beta_0} + \hat{\beta_1}x
    • y^\hat{y}: predicted value of dependent variable for given value of independent variable xx

Interpretation of regression components

  • Slope β1^\hat{\beta_1} represents change in dependent variable yy for one-unit change in independent variable xx
    • Positive slope indicates positive linear relationship (weight and height)
    • Negative slope indicates negative linear relationship (age and reaction time)
  • Y-intercept β0^\hat{\beta_0} represents value of dependent variable yy when independent variable xx equals zero
    • May not have meaningful interpretation if range of xx values does not include zero (negative age)

Predictions using regression equation

  • Substitute given value of independent variable xx into equation: y^=β0^+β1^x\hat{y} = \hat{\beta_0} + \hat{\beta_1}x
  • Resulting value y^\hat{y} is predicted value of dependent variable for given value of independent variable
  • Predictions subject to uncertainty
    • Accuracy depends on strength of linear relationship and quality of model fit (correlation coefficient, residual analysis)