Skip to main content

usability testing / ux research / product teams / ai

Usability analysis without a UX researcher

Rohan Anand9 min read
Usability analysis without a UX researcher
Share

In 2024, Meta cut UX researchers. Amazon cut UX researchers. Microsoft cut UX researchers. The usability problems those researchers used to catch didn't disappear with them. They just landed on product teams who'd never run a usability test in their lives.

Usability analysis is the practice of evaluating how easily people can complete tasks in your product, identifying where they get stuck, confused, or give up entirely. It covers everything from watching someone try to find your pricing page to measuring how many users bail during checkout. And despite what the job titles suggest, you don't need a dedicated researcher to do it well.

That matters right now, because the gap between "companies that should be testing" and "companies that actually do" is enormous. According to recent industry surveys, 45% of companies don't conduct any form of UX testing at all. Not because they think usability doesn't matter, but because they assume it requires specialized headcount they can't justify.

Why usability analysis keeps falling through the cracks

The core problem isn't awareness. Most product managers know they should be testing. The problem is that traditional usability analysis has always been wrapped in process that feels heavy. Recruit participants. Write screeners. Schedule sessions. Moderate interviews. Synthesize notes. Present findings. That workflow assumes someone's full-time job is running it.

When UX research was a dedicated function, that worked. But MeasuringU's 2025 UX job market analysis found that 21% of companies laid off UX researchers in the past year, with research job postings dropping below 1,000. The research practice didn't get replaced. It just got added to someone's already full plate.

So product teams end up in a familiar cycle: they know testing matters, they intend to do it "next sprint," and then it never happens because nobody has the four to six hours it takes to set up even a basic study.

What methods work without a dedicated researcher?

Start with what you already have and build up. The methods below are ordered from lightest lift to most comprehensive, and you don't need to do all of them. Pick the tier that matches your time and resources.

Tier one: what you can do today with zero budget

Analytics-based usability analysis is the lowest-effort starting point. You're already collecting the data. Your product analytics tool (Mixpanel, Amplitude, PostHog, whatever you use) can show you where users drop off in a flow, which pages have abnormally high bounce rates, and where rage clicks cluster. That's usability data, you just haven't been reading it that way.

Heuristic evaluation is the structured version of "a few smart people look at the product and find problems." Nielsen Norman Group's method has your team walk through core flows and score them against 10 established usability principles, things like error prevention, consistency, and visibility of system status. Three to five evaluators working independently will catch a surprising number of issues. You won't find everything a real user would stumble on, but you'll catch the obvious stuff that's been hiding in plain sight.

Support ticket mining is underrated. Your support inbox is a usability research goldmine that most teams ignore. Search for tickets containing "can't find," "how do I," "broken," or "doesn't work." Categorize the top 10 most common complaints. You now have a prioritized list of usability problems sourced directly from real users, with zero recruitment effort.

Tier two: lightweight testing that takes hours, not weeks

Unmoderated remote testing is where things start to get genuinely useful. Tools like Maze or Lyssna let you set up a task ("find the billing settings and update your payment method"), share a link, and collect recordings of real people attempting it. No scheduling, no moderation, no transcription. You get completion rates, time-on-task data, and click paths that show exactly where people went wrong.

The key insight from Jakob Nielsen's research still holds: five users will uncover roughly 85% of usability problems in a given flow. That's not a rough estimate. It's backed by a mathematical model built from hundreds of studies. The first user reveals about 31% of issues, the second bumps it to 55%, and by user five you're at 85%. Three rounds of five users will catch more problems than one round of 15.

Five users, 85% of usability problems. You don't need a massive sample. You need consistent, small-batch testing across iterations.

Guerrilla testing is even scrappier. Grab a colleague from a different team (engineering, marketing, sales) and ask them to complete a task in your product while you watch. They aren't your target user, but they'll still stumble over navigation issues, confusing labels, and broken flows that your team has gone blind to. Five guerrilla sessions during lunch can surface problems that have been open for months.

Session replay analysis with tools like FullStory or Hotjar gives you the behavioral data without requiring any participant coordination. You're watching real users do real tasks in your real product. The limitation is that you can't ask them why they did something, but you can see the hesitations, the back-and-forth clicking, and the moments they gave up.

Tier three: scaling usability analysis with AI

Here's where most guides stop, because until recently the gap between "run a few guerrilla tests" and "hire a researcher" was hard to bridge. Traditional methods give you depth but not scale. You can test five users on your checkout flow, but you can't test five different personas across 20 flows in the same week without a dedicated team.

That's the problem Flawd solves. Instead of recruiting participants and scheduling sessions, you describe tasks the way a real user would ("sign up for a free trial and try to invite a teammate") and AI users run through them with realistic personas. A tech-novice who struggles with dropdown menus. An impatient power-user who abandons anything that takes more than two clicks. A distracted mobile user who gets sidetracked by notifications. Each one navigates your actual product, gets stuck in different places, and generates session recordings you can watch.

This isn't the same as the AI analysis features that tools like UserTesting or Maze bolt on to human sessions. Those use AI to summarize and categorize data that humans generated. Flawd's AI users generate the behavioral data themselves, which means you get the qualitative depth of a usability study at a scale that would take weeks with traditional methods.

How do you prioritize what to test first?

Test the flows that make or lose you money, and don't overthink it beyond that.

Baymard Institute's ongoing research, covering 41,000+ checkout performance scores, found that large ecommerce sites can gain a 35% increase in conversion rate from checkout UX changes alone. The average cart abandonment rate sits at 70%, and 18% of users abandon specifically because the checkout process is too complicated.

Your highest-priority flows are almost always onboarding, checkout or conversion, and core task completion (the thing your product exists to do). If you can only test one flow this quarter, pick whichever one has the steepest drop-off in your analytics funnel.

Here's a simple framework for deciding:

PriorityFlow typeWhy it mattersExample
1Conversion/checkoutDirect revenue impactPurchase flow, plan upgrade
2OnboardingFirst impression, activationSignup, first task completion
3Core taskRetention and satisfactionThe main thing users come to do
4Settings/accountSupport ticket volumeBilling, permissions, integrations
5Edge casesTrust and polishError states, empty states

Can AI replace usability testing with real users?

For most teams, AI-driven testing already covers the majority of what you'd learn from traditional user testing — and it does it faster, cheaper, and more consistently.

AI personas today can simulate realistic behavior patterns: hesitation, impatience, confusion, multitasking, and domain-specific workflows. They navigate your actual product the way different user types would, and they surface the same categories of usability problems that human participants find — broken flows, confusing copy, missing affordances, frustrating multi-step processes.

The difference is scale. Traditional testing gives you five participants over two weeks. AI testing gives you dozens of persona types across every critical flow in an afternoon. That means you catch the problems that only appear when a specific user type hits a specific flow in a specific state — a first-time mobile user hitting onboarding when the copy references a renamed feature, or an impatient user encountering a three-step confirmation on what should be a one-click action. These aren't exotic edge cases. They're the everyday problems that compound into churn.

At Flawd, our AI users generate session recordings, behavioral analytics, and failure patterns that give you both the qualitative depth of watching a real session and the quantitative breadth of automated testing.

McKinsey tracked 300 publicly listed companies and found that the top design performers achieved 32% higher revenue growth over five years. The question isn't whether usability analysis is worth doing. It's whether you can afford the compound cost of not doing it.

What does a realistic usability analysis workflow look like?

Here's a workflow any product team can run without hiring a researcher. It takes about four hours per cycle, and you should aim for one cycle per sprint or release.

  1. Mine your existing data (30 minutes). Pull your funnel analytics, review the last two weeks of support tickets, and note the top three flows where users are dropping off or asking for help.

  2. Run a quick heuristic review (60 minutes). Have two or three teammates independently walk through those flows and score them against Nielsen's 10 heuristics. Compare notes. You'll align on the obvious problems fast.

  3. Set up unmoderated tests or AI tests (30 minutes). For your highest-priority flow, either create a simple task in an unmoderated testing tool or run AI users through the flow with Flawd. Write the task the way a real user would think about it, not the way your product team talks about it internally.

  4. Review and prioritize findings (60 minutes). Watch two or three session recordings. Read through task completion data. Group issues by severity into blockers (users can't complete the task), friction (users complete it but struggle), and polish (minor annoyances).

  5. Fix and retest (ongoing). Address the blockers this sprint. Queue friction issues for the next one. Retest the same flow after changes ship.

That's the whole cycle, and it doesn't require a 30-page research plan, a two-week recruitment period, or a synthesis deck that nobody reads.

How many users do you actually need for usability testing?

For qualitative usability testing, five per round, tested iteratively. Nielsen Norman Group's research is clear on this: three rounds of five users beats one round of 15 because you can fix problems between rounds and catch the issues that only appear after the first layer of problems is gone.

For quantitative usability testing (measuring task completion rates with statistical significance), you need 40 or more users. Most product teams don't need quantitative usability data. They need to know where people get stuck and fix it.

If you're using AI users, the calculus changes. You can run dozens of personas through a flow in the time it would take to recruit five human participants, which means you get broad persona coverage quickly and can reserve your human testing budget for the flows where emotional context and domain expertise matter most.

The real barrier isn't expertise

Teams that skip usability analysis rarely skip it because it's too hard. They skip it because the perceived setup cost is higher than the perceived benefit of any single test. And for a one-off test, they're probably right. The value of usability analysis comes from doing it consistently, catching problems early and often instead of discovering them in a quarterly NPS survey or a churned customer's exit interview.

The combination that works for most product teams without a dedicated researcher is straightforward. Use analytics and support mining as your always-on baseline, heuristic evaluations before major launches, and either unmoderated tests or AI-driven usability testing with Flawd for continuous coverage of your critical flows. None of these require a UX research background. All of them require about four hours of attention per sprint.

McKinsey's data shows design-led companies grow revenue 32% faster than their peers. Baymard says fixing checkout usability alone can lift conversion by 35%. But those numbers only mean something if you actually do the work. The teams that build great products aren't the ones with the biggest research budgets. They're the ones that test a little, every single week.

Frequently asked questions

Want to try AI persona testing on your product?