Study smarter with Fiveable
Get study guides, practice questions, and cheatsheets for all your subjects. Join 500,000+ students with a 96% pass rate.
Get Started
Why This Matters
User research isn't just a checkbox in the design process—it's the foundation that separates intuition-driven design from evidence-based strategy. You're being tested on your ability to select the right methodology for the right research question, understand when qualitative depth beats quantitative breadth, and recognize how different methods reveal different types of user insights. The methodologies here demonstrate core principles like triangulation, ecological validity, iterative validation, and participatory design thinking.
Don't just memorize what each method does. Know when to deploy it, what type of data it generates, and how it connects to strategic design decisions. An exam question won't ask you to define card sorting—it'll ask you to justify why card sorting beats interviews for an information architecture problem, or when A/B testing fails to capture the insights you actually need.
Immersive & Contextual Methods
These methodologies prioritize ecological validity—understanding users in their real environments rather than artificial lab settings. The underlying principle is that user behavior changes when removed from context, so researchers must go to the user.
Ethnographic Research
- Deep immersion in natural environments—researchers observe users over extended periods to uncover behaviors users themselves may not articulate
- Cultural and contextual factors surface through this method, revealing the why behind user actions that surveys miss entirely
- Rich qualitative data makes this ideal for early discovery phases when you don't yet know what questions to ask
Contextual Inquiry
- Hybrid observation-interview approach—researchers watch users perform tasks while asking clarifying questions in real-time
- Master-apprentice model positions the user as expert, with the researcher learning their workflow and mental models
- Task-level insights emerge that reveal workarounds, pain points, and unspoken needs users have normalized
Diary Studies
- Longitudinal self-reporting—participants document experiences over days or weeks, capturing data researchers couldn't observe directly
- Temporal patterns in behavior, emotion, and context become visible through this extended timeframe
- Ecological validity remains high since users record in-the-moment experiences rather than reconstructing from memory
Compare: Ethnographic research vs. Contextual inquiry—both prioritize real-world context, but ethnography is passive observation over time while contextual inquiry is active questioning during specific tasks. Use ethnography for discovery; use contextual inquiry when you've identified specific workflows to investigate.
Direct Elicitation Methods
These methods ask users directly about their experiences, needs, and preferences. The core principle is that users possess knowledge about their own needs—but researchers must structure the elicitation carefully to extract actionable insights rather than surface-level opinions.
User Interviews
- One-on-one conversations allow deep exploration of individual perspectives, motivations, and mental models
- Semi-structured formats balance consistency across participants with flexibility to pursue unexpected insights
- Rapport building enables disclosure of frustrations and workarounds users might not share in group settings
Surveys and Questionnaires
- Quantitative reach at scale—collect data from hundreds or thousands of users to identify patterns and measure prevalence
- Standardized questions enable statistical analysis and comparison across user segments or time periods
- Best for validation, not discovery—use surveys to test hypotheses generated through qualitative methods
Focus Groups
- Group dynamics generate emergent insights—participants build on each other's ideas and challenge assumptions in real-time
- Social context reveals how users discuss and frame their needs, useful for understanding shared mental models
- Groupthink risk requires skilled facilitation to prevent dominant voices from skewing results
Compare: User interviews vs. Focus groups—interviews reveal individual depth and private frustrations; focus groups reveal social dynamics and shared language. If an FRQ asks about sensitive topics or individual workflows, interviews win. For understanding how a community thinks about a product category, focus groups provide richer context.
Behavioral Testing Methods
These methodologies observe what users actually do rather than what they say they do. The principle here is that self-reported preferences often diverge from actual behavior—so direct observation of task performance yields more reliable design guidance.
Usability Testing
- Task-based observation—watch real users attempt specific tasks to identify friction points and failure modes
- Think-aloud protocol captures user reasoning and confusion in real-time, revealing why interactions fail
- Iterative validation makes this essential for testing prototypes before development investment
Remote User Testing
- Distributed participation enables testing with geographically diverse users in their actual environments
- Unmoderated options scale efficiently, though moderated remote sessions preserve the ability to probe deeper
- Authentic context captures real-world distractions and device configurations that lab settings miss
A/B Testing
- Controlled experimentation—randomly assign users to variants and measure behavioral differences statistically
- Quantitative optimization identifies which design performs better on specific metrics like conversion or engagement
- Limitation awareness is critical—A/B testing tells you what works better but not why, and only tests options you've already conceived
Eye Tracking
- Visual attention mapping—precisely measures where users look, for how long, and in what sequence
- Unconscious behavior is captured that users cannot self-report, revealing true attention patterns
- Heat maps and gaze plots inform visual hierarchy decisions and identify elements users overlook entirely
Compare: Usability testing vs. A/B testing—usability testing is qualitative and diagnostic (why do users struggle?), while A/B testing is quantitative and evaluative (which version performs better?). Use usability testing to identify problems; use A/B testing to validate solutions at scale.
These methods specifically address how users organize, categorize, and navigate information. The underlying principle is that designers' mental models often differ from users'—so IA decisions must be grounded in user cognition, not internal logic.
Card Sorting
- User-generated categorization—participants group content items in ways that make sense to them, revealing natural mental models
- Open vs. closed variants serve different purposes: open sorting discovers categories; closed sorting validates proposed structures
- Navigation and labeling decisions should flow directly from card sorting results to match user expectations
Heuristic Evaluation
- Expert inspection method—trained evaluators assess interfaces against established usability principles (like Nielsen's heuristics)
- Cost-efficient identification of obvious usability violations without recruiting users
- Complement, not replacement for user testing—experts catch different issues than real users do
Compare: Card sorting vs. Heuristic evaluation—card sorting captures user mental models for information organization, while heuristic evaluation applies expert knowledge of usability principles. Card sorting answers "how do users think about this content?" Heuristic evaluation answers "does this interface follow established best practices?"
Synthesis & Modeling Methods
These methodologies transform raw research data into actionable design tools. The principle is that insights must be synthesized into formats that guide ongoing design decisions—research findings alone don't drive action.
Personas and User Scenarios
- Archetypal user representations—personas synthesize research into memorable characters that keep user needs present in design discussions
- Scenario narratives describe how personas interact with products in specific contexts, grounding abstract features in concrete use cases
- Stakeholder alignment tool that prevents teams from designing for themselves rather than actual users
Journey Mapping
- End-to-end experience visualization—maps every touchpoint, emotion, and action across the full user journey
- Pain point identification becomes systematic when the entire experience is visible in one artifact
- Cross-functional alignment helps teams see how their piece fits into the user's holistic experience
Participatory Design
- Users as co-designers—involve users directly in generating and evaluating design solutions, not just providing feedback
- Ownership and buy-in increase when users shape the solutions they'll eventually use
- Power dynamics shift requires careful facilitation to ensure user input genuinely influences outcomes
Compare: Personas vs. Journey maps—personas represent who your users are; journey maps represent what they experience over time. Both are synthesis tools, but personas guide feature prioritization while journey maps guide experience optimization across touchpoints.
Quick Reference Table
|
| Ecological validity (real-world context) | Ethnographic research, Contextual inquiry, Diary studies |
| Quantitative measurement at scale | Surveys, A/B testing |
| Qualitative depth and discovery | User interviews, Ethnographic research, Focus groups |
| Behavioral observation (what users do) | Usability testing, Eye tracking, A/B testing |
| Information architecture decisions | Card sorting, Heuristic evaluation |
| Longitudinal/temporal insights | Diary studies, Journey mapping |
| Synthesis and modeling | Personas, Journey maps, User scenarios |
| Collaborative/participatory approaches | Participatory design, Contextual inquiry |
Self-Check Questions
-
You need to understand why users abandon a checkout flow, but you don't know where the friction occurs. Which two methodologies would you combine, and why does each contribute something the other can't?
-
Compare and contrast ethnographic research and contextual inquiry. When would you choose immersive observation over the hybrid interview-observation approach?
-
A stakeholder wants to run an A/B test to decide between two navigation structures. What's the limitation of this approach, and which methodology should precede it?
-
Which three methodologies are best suited for discovery phases when you don't yet know what questions to ask? What do they have in common?
-
Your team has completed user interviews and usability testing. Now you need to synthesize findings into artifacts that will guide design decisions for the next six months. Which two synthesis methods would you create, and what different purposes do they serve?