Study smarter with Fiveable
Get study guides, practice questions, and cheatsheets for all your subjects. Join 500,000+ students with a 96% pass rate.
Regression models form the backbone of statistical prediction and machine learning—and you're being tested on knowing when to use each type, not just what they are. The exam will challenge you to match real-world scenarios to appropriate models, understand the tradeoffs between model complexity and interpretability, and recognize when techniques like regularization, variable selection, and nonlinear transformations become necessary. These aren't just formulas to memorize; they're tools that solve fundamentally different problems.
The key insight connecting all regression models is the bias-variance tradeoff. Simple models (like linear regression) may underfit complex data, while flexible models (like high-degree polynomials) risk overfitting. Regularization techniques exist precisely to navigate this tension. When you encounter a regression question, don't just ask "what's the formula?"—ask "what problem does this model solve, and what assumptions does it make?" That's how you'll tackle both multiple-choice questions and FRQs with confidence.
These models assume that relationships between predictors and outcomes can be captured with linear combinations. The core assumption is that the effect of each predictor is additive and proportional.
Compare: Linear Regression vs. Multiple Linear Regression—both assume linearity and use least squares, but multiple regression introduces multicollinearity concerns and partial effect interpretation. FRQs often ask you to explain what "holding other variables constant" means practically.
When you have many predictors or correlated features, standard regression can overfit or produce unstable estimates. Regularization adds a penalty term to the loss function, shrinking coefficients toward zero.
Compare: Ridge vs. Lasso—both regularize, but ridge keeps all predictors (just shrunk) while lasso performs selection. If an FRQ asks about feature selection with many predictors, lasso is your go-to answer. Ridge is better when you believe all predictors contribute.
When the true relationship between variables curves, bends, or follows a nonlinear pattern, these models introduce flexibility through polynomial terms or general nonlinear functions.
Compare: Polynomial vs. Nonlinear Regression—polynomial regression is technically linear in parameters (just transform ), while true nonlinear regression has parameters inside nonlinear functions. Polynomial is easier to fit; nonlinear requires more computational effort but can match theory-driven functional forms.
Not all outcomes are continuous. When your dependent variable is categorical or represents counts, you need models designed for those distributions.
Compare: Logistic vs. Poisson Regression—logistic handles binary yes/no outcomes, Poisson handles "how many" count outcomes. Both use maximum likelihood and link functions, but they model fundamentally different types of dependent variables. Know which outcome type triggers which model.
Sometimes the goal isn't just fitting data—it's finding the simplest adequate model. These techniques help you decide which predictors to include.
Compare: Stepwise vs. Lasso—both perform variable selection, but stepwise uses discrete add/remove decisions while lasso uses continuous shrinkage. Lasso is generally preferred in modern practice because it's less prone to overfitting and handles correlated predictors better.
Standard regression assumes observations are independent. When data points are ordered in time, you need models that account for temporal structure.
Compare: Time Series Regression vs. Standard Multiple Regression—both can have multiple predictors, but time series regression explicitly models temporal dependencies. Ignoring autocorrelation in time data leads to incorrect standard errors and misleading inference.
| Concept | Best Examples |
|---|---|
| Linear relationships | Linear Regression, Multiple Linear Regression |
| Regularization to prevent overfitting | Ridge Regression, Lasso Regression |
| Variable/feature selection | Lasso Regression, Stepwise Regression |
| Handling multicollinearity | Ridge Regression, Lasso Regression |
| Nonlinear patterns | Polynomial Regression, Nonlinear Regression |
| Binary/categorical outcomes | Logistic Regression |
| Count data | Poisson Regression |
| Temporal dependencies | Time Series Regression |
You have 50 predictors but suspect only 5-10 are truly relevant. Which regression technique performs automatic variable selection, and how does it differ from ridge regression?
Compare and contrast polynomial regression and nonlinear regression. Why is polynomial regression considered "linear" even though it can fit curves?
A researcher is modeling the number of customer complaints per day. Why would Poisson regression be more appropriate than linear regression, and what assumption should they check?
If an FRQ presents a dataset with highly correlated predictors and asks you to build a stable predictive model, which two regularization approaches would you consider, and what's the key tradeoff between them?
Explain why ignoring temporal autocorrelation when analyzing time-ordered data leads to problems, even if you use multiple linear regression with relevant predictors.