upgrade
upgrade

🐛Biostatistics

Types of Study Designs

Study smarter with Fiveable

Get study guides, practice questions, and cheatsheets for all your subjects. Join 500,000+ students with a 96% pass rate.

Get Started

Why This Matters

In biostatistics, you're not just learning a list of study designs—you're learning a toolkit for evaluating evidence. Every research question demands a specific approach, and understanding why researchers choose RCTs over cohort studies, or when case-control designs outperform cross-sectional ones, is fundamental to interpreting health data. Exams will test your ability to match research questions to appropriate designs, identify sources of bias, and evaluate the strength of causal claims.

The key concepts here are directionality (prospective vs. retrospective), researcher control (experimental vs. observational), and evidence hierarchy (which designs support causal inference). Don't just memorize definitions—know what each design can and cannot tell us, and be ready to explain why a researcher would choose one over another.


Experimental Designs: The Gold Standard for Causation

Experimental studies give researchers control over exposures, allowing them to isolate cause-and-effect relationships. By manipulating variables and randomizing participants, these designs minimize confounding and bias.

Randomized Controlled Trials (RCTs)

  • Random assignment to treatment or control groups—this eliminates selection bias and balances confounders across groups
  • Gold standard for causal inference because the only systematic difference between groups should be the intervention itself
  • Requires ethical approval and informed consent, limiting what can be studied (you can't randomly assign people to smoke)

Experimental Studies (General)

  • Researcher manipulates at least one independent variable to observe effects on outcomes
  • Includes RCTs, quasi-experiments, and laboratory studies—any design where intervention is controlled
  • Strongest internal validity when properly designed, but may sacrifice external validity (real-world applicability)

Compare: RCTs vs. Quasi-Experiments—both involve intervention, but RCTs use randomization while quasi-experiments do not. If an exam asks about causation with the strongest evidence, RCTs are your answer; quasi-experiments are the fallback when randomization isn't feasible.


Observational Designs: Studying Without Intervening

When researchers cannot or should not manipulate variables, observational designs allow them to study associations in natural settings. The tradeoff is reduced control over confounding variables.

Cohort Studies

  • Follows a defined group over time to compare outcomes between exposed and unexposed individuals
  • Can be prospective or retrospective—prospective designs collect data as events unfold; retrospective designs use existing records
  • Calculates relative risk directly, making it ideal for studying incidence and multiple outcomes from a single exposure

Case-Control Studies

  • Starts with outcome, looks backward for exposures—cases (with disease) are compared to controls (without)
  • Highly efficient for rare diseases because you don't need to follow thousands of people waiting for outcomes to occur
  • Calculates odds ratios, not relative risk—a key distinction for exam questions about appropriate measures of association

Cross-Sectional Studies

  • Snapshot of a population at one point in time—measures exposure and outcome simultaneously
  • Calculates prevalence, not incidence—useful for public health planning and hypothesis generation
  • Cannot establish temporality, meaning you can't determine whether exposure preceded outcome (chicken-or-egg problem)

Compare: Cohort vs. Case-Control—both are analytic observational designs, but cohort studies follow exposure → outcome (prospective logic) while case-control studies work backward from outcome → exposure. FRQs often ask which design is better for rare diseases (case-control) vs. rare exposures (cohort).


Temporal Direction: Prospective vs. Retrospective

The direction of data collection fundamentally affects bias and data quality. Prospective designs collect data as events happen; retrospective designs rely on existing records or memory.

Prospective Studies

  • Data collected forward in time from exposure to outcome—researchers define measurements before outcomes occur
  • Minimizes recall bias because participants report exposures before knowing their outcome status
  • More expensive and time-consuming, especially for diseases with long latency periods

Retrospective Studies

  • Uses historical data to assess past exposures and outcomes—faster and cheaper than prospective approaches
  • Vulnerable to recall bias and incomplete records—participants may misremember exposures, especially after diagnosis
  • Essential for rare outcomes where waiting for prospective data would be impractical

Longitudinal Studies

  • Repeated measurements over extended time periods—captures change, trends, and temporal sequences
  • Can be observational or experimental, prospective or retrospective—the defining feature is multiple time points
  • Best for studying disease progression and developmental trajectories (how does health status evolve?)

Compare: Prospective vs. Retrospective Cohort—both track exposure → outcome, but prospective cohorts collect data in real-time while retrospective cohorts use existing records. Prospective designs have better data quality; retrospective designs are faster and cheaper.


Synthesizing Evidence: Combining Multiple Studies

When individual studies provide conflicting or limited evidence, systematic approaches can synthesize findings across research. Meta-analysis uses statistical methods to pool results and increase precision.

Meta-Analyses

  • Statistically combines results from multiple studies on the same question—treats each study as a data point
  • Increases statistical power to detect effects that individual studies might miss due to small sample sizes
  • Quality depends entirely on included studies—"garbage in, garbage out" applies; heterogeneity between studies must be assessed

Compare: Meta-Analysis vs. Systematic Review—both synthesize multiple studies, but systematic reviews summarize findings qualitatively while meta-analyses pool data quantitatively. Meta-analyses produce a single effect estimate; systematic reviews may not.


Quick Reference Table

ConceptBest Examples
Strongest causal evidenceRCTs, Experimental Studies
Efficient for rare diseasesCase-Control Studies
Calculates incidence/relative riskCohort Studies (prospective)
Calculates prevalenceCross-Sectional Studies
Minimizes recall biasProspective Studies
Fastest/cheapest approachRetrospective Studies, Case-Control
Studies change over timeLongitudinal Studies, Cohort Studies
Synthesizes multiple studiesMeta-Analyses

Self-Check Questions

  1. A researcher wants to study risk factors for a rare childhood cancer. Which study design is most efficient, and why can't they use an RCT?

  2. Compare cohort studies and case-control studies: which calculates relative risk directly, and which calculates odds ratios? When does the odds ratio approximate relative risk?

  3. A cross-sectional study finds that people who exercise have lower rates of depression. Why can't the researchers conclude that exercise prevents depression?

  4. What distinguishes a prospective cohort study from a retrospective cohort study? Which has better protection against recall bias, and why?

  5. If three RCTs on the same drug show conflicting results, what study design could help resolve the discrepancy, and what limitation should you consider when interpreting its findings?