User Research Methods: 13 Techniques and How to Choose (2026)
Johannes
Co-Founder
11 Minutes
May 3rd, 2026
Most product teams default to the same 1-2 familiar methods regardless of what question they're actually trying to answer. The result is attitudinal data when you need behavioral evidence, or qualitative depth when you need statistical breadth. This guide covers 13 user research methods, organized by three practical dimensions, with a decision framework for choosing between them and clear guidance on how to run each one.
The 3 Dimensions Every Research Method Falls Into
Before picking a method, understand three dimensions that define how any research technique works. Mapping your method against these helps you match approach to research question.
Dimension 1: Attitudinal vs. Behavioral
- Attitudinal research captures what people say -- their opinions, preferences, and self-reported behaviors.
- Behavioral research captures what people actually do, independent of what they say.
The gap between these two is often where the most important findings live. A user might say they prioritize data security but still reuse the same password across every account.
Dimension 2: Qualitative vs. Quantitative
- Qualitative methods answer why -- they generate hypotheses, surface unexpected patterns, and add context to numbers.
- Quantitative methods answer how many and how much -- they validate hypotheses, measure magnitude, and track change over time.
Neither is superior. The research question determines which you need.
Dimension 3: Context of Use
Where does the research happen? The spectrum runs from natural use in the field (contextual inquiry, diary studies) to scripted lab scenarios (usability testing) to fully unmoderated remote sessions (unmoderated testing, session recording). Context affects what behaviors you observe and how far findings generalize.
Generative vs. Evaluative Research
A fourth dimension cuts across all three above: whether you're in generative or evaluative mode.
Generative research happens early. You're trying to understand the problem space -- who has this problem, why it matters, and what constraints shape behavior. Methods: user interviews, contextual inquiry, diary studies.
Evaluative research happens once there's something to test. You're checking whether a solution works and where it breaks. Methods: usability testing, A/B testing, tree testing.
The most common mistake is running evaluative research on an unvalidated problem. Testing a prototype before confirming you're solving the right problem produces confident wrong answers.
How to Choose the Right Method
| Research Question | Best Methods | Why |
|---|---|---|
| Why do users abandon the checkout? | User interviews, session recording | Need qualitative context for behavioral signal |
| Does version A or B convert better? | A/B testing | Quantitative comparison with statistical confidence |
| How do users navigate our information architecture? | Tree testing, card sorting | Tests mental models without visual design interference |
| What do users do between sessions? | Diary studies | Captures natural behavior over time |
| Does this prototype make sense? | Usability testing | Direct observation of task completion |
| What are users' top frustrations at scale? | Surveys | Quantifies sentiment across a large sample |
| How does our product fit into users' work? | Contextual inquiry | Observation in real environment |
| Where do users look first on a page? | Eye tracking | Measures attention, not self-report |
1. User Interviews
User interviews are structured one-on-one conversations designed to surface the reasoning behind user behavior. Unlike surveys, they capture nuance -- hesitation, contradiction, and context -- that closed questions miss.
Use them for generative research during discovery, or to investigate specific drop-offs that quantitative data has flagged but not explained.
How to run them well:
- Write a semi-structured guide, not a rigid script. You need flexibility to follow interesting threads.
- Ask about past behavior, not hypothetical futures. "Tell me about the last time you tried to accomplish X" is more reliable than "Would you use a feature that does Y?"
- Record sessions with consent so you can focus on listening rather than transcription.
- Recruit 6-8 participants per user segment. Patterns in qualitative research emerge quickly; more doesn't always mean better.
Common mistake: Asking leading questions that confirm what you already believe. Keep your guide open-ended and your framing neutral.
You can use Formbricks to run targeted screener surveys inside your product to recruit the right interview participants. Filter by recent behavior, plan type, or feature usage so you're always talking to the most relevant users.
2. Usability Testing
Usability testing assigns participants specific tasks to complete while researchers observe. The goal is to find where users get confused, stuck, or take unexpected paths -- not to confirm that your design works.
Nielsen and Landauer (1993) established the mathematical relationship between participant count and problems found. Their model showed that each additional participant reveals progressively fewer new issues, supporting small, frequent rounds of testing over large infrequent ones.
How to run it well:
- Design tasks around user goals, not features. "Find and export your monthly report" beats "click the export button."
- Use the think-aloud protocol: ask participants to narrate their reasoning as they work.
- Test early. A paper prototype with 5 users will surface more actionable issues per hour than repeated design reviews of finished screens.
- Run multiple rounds with iterations between them. One test is a snapshot; repeated testing is a learning loop.
Moderated vs. unmoderated:
| Type | Best For | Trade-offs |
|---|---|---|
| Moderated | Complex flows, early-stage prototypes | Time-intensive, smaller samples |
| Unmoderated | Established flows, specific task metrics | Less context, harder to probe follow-up questions |
Use Formbricks to launch a post-task survey immediately after users complete a key workflow. In-product surveys triggered at the moment of action capture more accurate sentiment than feedback requests sent hours later.
3. Surveys and Questionnaires
Surveys collect structured feedback from large groups through standardized questions. They are the most scalable user research method for quantifying sentiment, measuring change over time, and segmenting by user behavior or demographics. For guidance on how to get surveys in front of the right people, see our post on survey distribution methods.
The tradeoff is depth. Surveys tell you what users think, not why. Krosnick (1999) documented "satisficing" -- where respondents give acceptable-but-inaccurate answers to minimize cognitive effort -- making question design critical to data quality.
How to run them well:
- Keep them short. Every question should trace directly to a research objective.
- Use clear, neutral language. Loaded phrasing produces skewed distributions.
- Mix question types: rating scales for quantifiable sentiment, open-ended fields for context you can't anticipate.
- Time them to relevant moments. A satisfaction survey triggered immediately after a key interaction is more accurate than one sent days later.
When surveys fall short: Surveys can't capture actual behavior. Pair survey data with analytics, session recording, or direct observation when you need behavioral evidence.
Formbricks lets you embed targeted surveys directly in your product, triggered by specific user actions. Ask a single follow-up question right after a user encounters a friction point to capture context while it's still fresh.
4. Contextual Inquiry
Contextual inquiry is an ethnographic method where researchers observe users in their natural environment -- their actual workspace, home, or wherever they use the product -- while asking questions in real time.
It's the most accurate method for capturing how a product fits into a user's actual workflow. Lab-based research misses the interruptions, workarounds, and environmental constraints that shape real behavior.
How to run it well:
- Frame yourself as the apprentice, the user as the expert. This reduces social desirability bias and encourages natural behavior.
- Observe first, ask later. Note what you see; clarify at natural pauses rather than interrupting the flow.
- Document the full context: physical environment, surrounding tools, interruptions. These details often explain the why behind what you observe.
- Keep groups small. One researcher and one participant is standard; larger observer groups change the dynamic.
When to use it: Early generative research, especially when designing for complex workflows or professional contexts where the environment significantly shapes behavior.
5. A/B Testing
A/B testing compares two or more versions of a single element by randomly assigning users to each version and measuring a defined outcome metric. It removes subjective judgment from design decisions and lets behavior determine the result.
How to run it well:
- Test one variable at a time. Simultaneous changes make results uninterpretable.
- Define your success metric before running the test. Post-hoc metric selection is a common source of false positives.
- Calculate required sample size upfront using a power analysis. Stopping early because early results "look promising" undermines statistical validity.
- Run tests for at least one full business cycle to account for day-of-week behavior patterns.
Common mistake: Running A/B tests without qualitative context. Knowing version B outperformed version A tells you nothing about why, which limits your ability to generalize the learning.
6. Card Sorting
Card sorting asks participants to organize topics, features, or content into groups that make sense to them. It reveals how users mentally model your information architecture -- essential for designing navigation systems and content hierarchies.
Open vs. closed card sorting:
| Type | What Users Do | Best For |
|---|---|---|
| Open | Create their own categories | Discovering how users naturally group content |
| Closed | Sort cards into predefined categories | Validating whether your existing structure makes sense |
| Hybrid | Sort and then name groups | Balancing discovery with validation |
How to run it well:
- Use 30-60 items. Fewer produces limited data; more overwhelms participants.
- Ask participants to think aloud. The reasoning behind groupings is often more valuable than the groupings themselves.
- Analyze with a similarity matrix to find consensus across participants.
- Run open sorting first to discover structure, then closed sorting to validate your interpretation.
To understand why validated structure alone isn't enough, also read our post on avoiding the feature chaser fallacy.
7. Tree Testing
Tree testing evaluates how well users can navigate an information architecture by asking them to find specific items within a text-only hierarchy -- no visual design, just structure.
It's the complement to card sorting. Card sorting tells you how users expect content to be organized; tree testing tells you whether your actual structure supports successful navigation.
How to run it well:
- Use realistic task scenarios: "Find where you would go to update your billing email."
- Recruit 50+ participants. Tree testing lends itself to larger samples than moderated usability testing.
- Analyze success rate, time on task, and the paths taken when users fail. Failed paths reveal where your structure mismatches user expectations.
When to use it: Before committing to a navigation redesign, or after card sorting to confirm the structure you built actually works for your users.
8. Diary Studies
Diary studies ask participants to self-report their experiences, behaviors, and thoughts at regular intervals over days or weeks. They capture longitudinal behavior in natural context -- something no lab study can replicate.
Bolger, Davis, and Rafaeli (2003) identified diary methods as uniquely suited to capturing experience as it unfolds, rather than through retrospective recall, which is subject to memory distortion and peak-end effects.
How to run them well:
- Keep individual entries short. Long prompts produce lower compliance and less honest responses.
- Use time-based prompts (e.g., every evening) or event-triggered prompts (e.g., after each relevant use) depending on your research question.
- Recruit for commitment. Dropout is the biggest methodological risk in diary research.
- Follow up with a debrief interview after the study period to probe patterns from the entries.
When to use them: When you need to understand behavior over time -- how attitudes shift during onboarding, how a tool gets used across different contexts throughout the week, or how emotional engagement with a product evolves.
9. Session Recording
Session recording captures video of real user sessions -- screen behavior, mouse movement, scroll depth, and in some setups face or voice -- as users interact with a product in their natural environment.
Unlike moderated testing, session recording happens without a researcher present. This reduces observation effects but removes the ability to ask clarifying questions in real time.
How to use it well:
- Filter for sessions around specific events: rage clicks, form abandonment, repeated errors. Random session browsing is inefficient.
- Look for patterns across sessions, not isolated anomalies. One user struggling is a data point; five users struggling at the same place is a finding.
- Combine with heatmaps and click maps for a spatial view of where attention concentrates.
- Obtain informed consent and handle session data according to applicable privacy regulations.
10. User Journey Mapping
User journey mapping visualizes the complete sequence of interactions a user has with a product over time, including their goals, actions, and emotional states at each stage.
Its primary value is alignment, not discovery. Journey maps synthesize existing research into a format that cross-functional teams can use to build shared understanding and prioritize where to improve.
How to run it well:
- Build it from research data, not assumptions. User interviews, surveys, and analytics should supply the substance.
- Include emotional highs and lows at each stage. The emotional layer is what separates a useful map from a process diagram.
- Keep it to one primary persona per map. Blending multiple personas produces a fictional average user who doesn't represent anyone.
- Revisit and update maps as the product evolves. A static journey map becomes misleading quickly.
Use Formbricks to trigger in-product surveys at specific touchpoints identified on your journey map. If your map shows a potential drop-off during onboarding, a targeted survey at that exact step can surface what's blocking users before they leave.
11. Persona Development
Personas are research-grounded character sketches representing distinct user types. They compress patterns from user research into memorable references that help teams make consistent decisions about who they're designing for.
The risk is confusing fictional personas with real user data. A persona built on assumptions reflects the team's biases. A persona built from interviews and behavioral data is a genuine design tool.
How to run it well:
- Build personas from primary research: interview transcripts, behavioral segmentation from analytics, survey data.
- Focus on goals and behaviors, not demographics. What a persona is trying to accomplish matters more than their age or job title.
- Create 3-5 primary personas. More than that and teams can't keep them in working memory.
- Include direct quotes from research participants to ground each persona in real language.
Use recurring Formbricks surveys to continuously validate your personas. A short quarterly survey asking about user goals or recent challenges provides fresh data to keep personas accurate as your user base evolves.
12. Eye Tracking
Eye tracking measures where users look on a page, in what sequence, and for how long. It reveals attentional patterns that self-report cannot capture -- users rarely know or accurately describe where they actually looked.
How to use it well:
- Use heatmaps to identify high-attention zones and overlooked areas.
- Focus on specific hypotheses. "Does our primary CTA get attention above the fold?" is testable. "How do users read our homepage?" is too broad to produce actionable findings.
- Pair with usability testing. Eye tracking data without behavioral context often produces impressive visualizations but limited practical insight.
- Lab-based eye tracking requires specialized hardware; webcam-based alternatives have improved but remain less accurate for precision tasks.
When to use it: For specific questions about visual hierarchy, attention to key UI elements, or how users scan content-heavy pages.
13. Analytics and Behavioral Data Analysis
Product analytics measures what users do at scale -- click patterns, session duration, feature adoption, conversion rates, and user flows. It is the primary source of behavioral data that doesn't require researcher intervention.
How to use it well:
- Define actionable metrics tied to user success. Conversion rates and task completion rates tell you more than page views.
- Segment by meaningful cohorts: new vs. returning users, acquisition channel, plan tier. Aggregate data often obscures the patterns that matter.
- Treat analytics as a problem identifier, not a problem explainer. A drop-off in your funnel is the start of a research question -- not the answer.
Combining with qualitative methods: Quantitative data shows what is happening. Qualitative methods explain why. Analytics without user research produces optimization without understanding.
Use Formbricks to trigger in-app surveys based on specific behavioral events. If a user visits your pricing page three times without upgrading, automatically surface a one-question survey -- "Is anything unclear about our pricing?" -- to convert behavioral signal into qualitative context.
Full Method Comparison Matrix
| Method | Type | Phase | Sample Size | Time to Insights | Best For |
|---|---|---|---|---|---|
| User Interviews | Qualitative / Attitudinal | Generative | 6-12 | 1-2 weeks | Understanding motivations, discovering unknowns |
| Usability Testing | Qualitative / Behavioral | Evaluative | 5-10 | 1-2 weeks | Finding friction in specific flows |
| Surveys | Quantitative / Attitudinal | Both | 100+ | Days | Scaling sentiment, validating hypotheses |
| Contextual Inquiry | Qualitative / Behavioral | Generative | 4-8 | 2-3 weeks | Understanding workflows in real context |
| A/B Testing | Quantitative / Behavioral | Evaluative | 1,000+ | Weeks-months | Comparing design variants with statistical confidence |
| Card Sorting | Quantitative / Attitudinal | Generative | 15-30 | 1-2 weeks | Designing information architecture |
| Tree Testing | Quantitative / Behavioral | Evaluative | 50+ | 1-2 weeks | Validating navigation structure |
| Diary Studies | Qualitative / Attitudinal | Generative | 10-20 | 2-4 weeks | Longitudinal experience in natural context |
| Session Recording | Quantitative / Behavioral | Evaluative | All sessions | Days | Identifying friction from real usage patterns |
| User Journey Mapping | Qualitative synthesis | Both | N/A | 1-2 weeks | Cross-team alignment, experience overview |
| Persona Development | Qualitative synthesis | Generative | N/A | 1-2 weeks | Communicating user types to the team |
| Eye Tracking | Quantitative / Behavioral | Evaluative | 15-30 | 1-2 weeks | Attention and visual hierarchy questions |
| Analytics | Quantitative / Behavioral | Both | All users | Days | Identifying patterns and problems at scale |
Building a Research Practice, Not Just Running Studies
The methods above are most effective when they're used in combination, not isolation. A common pattern:
- Analytics flags an unexpected drop-off in your onboarding flow.
- Session recording shows users repeatedly clicking a non-interactive element.
- User interviews reveal the mental model that explains why users expect that element to be clickable.
- Usability testing validates whether a redesign fixes the confusion.
- A/B testing confirms the fix improves conversion at scale.
Each method answers a different slice of the same question. Choosing the right one at each stage reduces waste and accelerates the time between research question and confident decision.
For teams new to user research, the best starting point is usually a combination of user interviews and a short contextual survey. This gives you both the depth of direct conversation and the scale of quantitative feedback without requiring significant tooling investment.
Ready to start collecting user feedback in context? Formbricks is an open-source survey and experience management platform built for product teams. Trigger targeted micro-surveys based on user behavior, segment responses by cohort, and maintain full data privacy. Get started with Formbricks.
Try Formbricks now
