Heuristic Evaluation: How to Conduct UX Audits When Working Solo

Heuristic Evaluation: How to Conduct UX Audits When Working Solo

By Paulina Contreras12 min read
📑 View contents

Initial Confession: The Elephant in the Room

Okay, let me be honest from the start.

If you Google “heuristic evaluation,” you’ll find tons of articles telling you that you need between 3 and 5 expert evaluators to identify 75-80% of usability problems. And they’re right. The literature says so, Nielsen says so, the papers confirm it.

The problem? I work in consulting. Solo.

Well, not entirely alone (hi Mauro 👋), but when I conduct heuristic evaluations for clients like OficinaVirtual or Mudango, I’m the one sitting in front of the screen reviewing flows and detecting problems. I don’t have a team of 5 UX researchers waiting to triangulate findings.

And you know what? That doesn’t invalidate the method. It just means you need to be more strategic about how you use it.

In this article, I’ll tell you how I adapt heuristic evaluation to the reality of working as an independent consultant in Chile, why I never use it in isolation, and how I combine it with other tools so the findings have the backing they deserve.

What is Heuristic Evaluation? (The Honest Version)

Heuristic evaluation is a usability inspection method where an expert (or several, in an ideal world) reviews an interface and compares it against a set of recognized design principles called heuristics.

The term comes from the Greek heuriskein, meaning “to discover.” And that’s exactly what we do: discover potential problems before real users get frustrated with them.

“Heuristic evaluation involves having a small set of evaluators examine the interface and judge its compliance with recognized usability principles.” — Jakob Nielsen

Why Is It Still Useful After 30 Years?

Nielsen’s 10 heuristics turned 31 in 2025, and they remain the de facto standard. Why? Because they’re not specific design rules that change with each new technology, but rather address fundamental human factors: memory capacity, the need for feedback, system control.

Human cognition doesn’t change at the same speed as technology.

What Heuristic Evaluation Is NOT

Before we continue, let’s clarify something important:

  • It’s not a usability test. You don’t observe real users doing tasks.
  • It’s not a cognitive walkthrough. Although they’re similar, a walkthrough simulates a new user’s step-by-step experience; heuristic evaluation assesses general principles.
  • It’s not a “pass/fail” checklist. This is one of the most common mistakes. Treating heuristics as a binary checklist ignores the method’s purpose: contextual diagnosis.

Nielsen’s 10 Heuristics: The Standard That Lives On

Although other frameworks exist (Shneiderman, Gerhardt-Powals, UX Quality Heuristics), Nielsen’s 10 heuristics remain the starting point for any evaluation. They’re the “lingua franca” of usability.

Here I’ll explain them with examples I’ve found in real projects:

1. Visibility of System Status

The system should keep users informed about what’s happening through appropriate feedback within a reasonable time.

Problem example: A quote form that shows no progress indicator. The user doesn’t know if they’re on step 2 of 3 or 2 of 10.

2. Match Between System and the Real World

The system should speak the users’ language, not technical jargon.

Problem example: A services site that uses “Peripheral Input Devices” when the user is looking for “Keyboards and Mice.” Or worse: using legal terms without explanation in a contracting flow.

3. User Control and Freedom

Users make mistakes and need clear “emergency exits.”

Problem example: A purchase flow where you can’t go back without losing everything you entered. Or a modal that can’t be closed without completing an action.

4. Consistency and Standards

The same term should mean the same thing throughout the site.

Problem example: An admin panel that uses “Dashboard” to refer to both the entire area AND a specific sub-page. This forces the user to learn two meanings for the same word.

5. Error Prevention

Better than a good error message is preventing the error from occurring.

Problem example: A date field that accepts any format but then fails validation. Better: a datepicker that only allows valid dates.

6. Recognition Rather Than Recall

Minimize memory load by making objects, actions, and options visible.

Problem example: A site that doesn’t show “recently viewed products” or recent searches. The user has to remember everything they did in previous sessions.

7. Flexibility and Efficiency of Use

The system should serve both novices and experts.

Problem example: Not offering keyboard shortcuts in a tool that frequent users use daily. Or conversely: overwhelming new users with advanced options.

8. Aesthetic and Minimalist Design

Every extra piece of information competes with relevant elements.

Problem example: A product page with so much information that the “Buy” button gets lost among banners, badges, and promotional text.

9. Help Users Recognize, Diagnose, and Recover from Errors

Error messages should be clear, indicate the problem, and suggest a solution.

Bad example: “SYNTAX ERROR” or “Error 500.”

Good example: “Password must have at least 8 characters. You’re missing 2.”

10. Help and Documentation

Although ideally the system shouldn’t need documentation, sometimes it’s necessary.

Problem example: A complex system with no tooltips, contextual guides, or help section. The user is left stranded.

Beyond Nielsen: Cugelman’s Behavioral Heuristics (AIM)

This is where it gets interesting (and where I differ from most articles on this topic).

In my last project with OficinaVirtual, I didn’t limit myself to Nielsen’s heuristics. I incorporated Brian Cugelman’s AIM framework, which comes from applied behavioral analysis.

What is the AIM Framework?

AIM stands for Attention, Interpretation, Motivation. It’s a framework developed by Cugelman and his team at AlterSpark that complements traditional heuristics with behavioral psychology principles.

The idea is simple but powerful:

  1. Attention: Does the design capture user attention at the right points?
  2. Interpretation: Does the user understand what they’re seeing? Does it make sense within their mental models?
  3. Motivation: Is there enough motivation to perform the desired action? Is it easy to execute?

This framework connects directly to the Fogg Model (B=MAP): a behavior occurs when there’s sufficient Motivation, Ability, and a Prompt at the right moment.

Why Do I Use It?

Because Nielsen’s heuristics tell you if something is well designed, but not necessarily if it will work for the business objective.

A button can perfectly comply with “Visibility of System Status” and still not generate conversions because the user isn’t motivated, or because the copy doesn’t connect with their mental models.

With AIM we evaluate:

  • Is the design capturing attention in the right place?
  • Does the language connect with the user’s mental models?
  • Is the process friction calibrated with the motivation level?

How I Do It in Practice: Real Cases

Let me tell you how I’ve applied this in real projects:

Case 1: OficinaVirtual (2025)

Context: Virtual office services startup that needed to review their new page’s sales flow before launch. They had only 2 weeks.

The challenge: Three simultaneous decisions on one screen (choose plan, choose location, choose term). Too much cognitive load.

My approach:

  1. Heuristic evaluation with Nielsen + AIM (Cugelman)
  2. Analytics and Search Console analysis to understand current behavior
  3. Benchmark of references (telcos, which handle plan presentation well)

What I found with traditional heuristics:

  • Violation of “Recognition rather than recall” (user had to mentally compare too many variables)
  • “Consistency” problems (the price table changed format based on selected term)

What I found with AIM:

  • Attention was scattered (the eye didn’t know where to look first)
  • Interpretation failed because the user’s mental model (comparing prices like a telco) didn’t match the presented structure
  • Motivation was lost in the friction of having to make 3 decisions together

Recommendation: Separate decisions into steps or use progressive disclosure. Highlight the best-selling plan as an anchor (a technique telcos use).

Case 2: Mudango (2023)

Context: Storage service that wanted to improve the quote flow through their Bot.

My approach:

  1. Heuristic evaluation of the web flow
  2. Funnel analysis in Analytics (from entry to conversion)
  3. In-depth interviews with customers (10 per country: Chile, Colombia, Mexico)
  4. Moderated usability testing with 5-7 people

Why I combined methods:

  • Heuristics gave me a quick diagnosis of obvious problems
  • Interviews revealed pain points I never would have found just looking at the interface
  • The Analytics funnel showed me WHERE users were dropping off
  • Usability tests CONFIRMED the hypotheses

Key learning: The heuristic evaluation was the initial filter that allowed me to prioritize which flows to test with users. Without it, I would have wasted time testing things that were obviously broken.

The Triangulation Method: Why I Never Use Heuristics Alone

Here comes the most important point of this article.

Heuristic evaluation is an initial quality filter, not a final diagnosis.

When you work solo (like me), you have to compensate for the lack of multiple evaluators with source triangulation. This means combining heuristic evaluation with at least 2-3 additional data sources.

My Typical Combo

Option A: When time is short (like OficinaVirtual)

  • Heuristic evaluation (Nielsen + AIM)
  • Analytics and Search Console analysis
  • Competitor/reference benchmark

Option B: When there’s more time and budget (like Mudango)

  • Heuristic evaluation
  • Internal data analysis (funnels, heatmaps, recordings)
  • User interviews
  • Usability testing

Why Does It Work?

Because each method gives you a piece of the puzzle:

Method Tells you…
Heuristic evaluation What MIGHT be wrong (hypothesis)
Analytics/Funnels WHERE users are dropping off
Interviews WHY they’re frustrated
Usability tests IF the problem is real and severe

Heuristics alone give you hypotheses. Triangulation gives you evidence.

And when I present results to a client, I don’t say “I think this is wrong.” I say “heuristics suggest a problem here, Analytics confirms there’s a 40% drop at this point, and interviews reveal that users don’t understand X.”

It’s the difference between opinion and evidence.

Step by Step: How to Execute Your Heuristic Evaluation

Okay, now the practical part. This is how I do a heuristic evaluation:

1. Define the Scope

Don’t try to evaluate the ENTIRE site. Focus on:

  • Critical flows (quote, purchase, registration)
  • Pages with the highest traffic or highest abandonment rates
  • New elements that are about to launch

2. Understand the Context

Before opening the interface:

  • Who are the target users?
  • What are the business objectives?
  • Are there defined personas/archetypes?

Without context, the evaluation becomes generic.

3. First Pass: Overview

I navigate the site without taking notes, just to understand the general flow. How does it feel? Where do I get lost? What frustrates me?

4. Second Pass: Systematic Evaluation

Now, with heuristics in hand (Nielsen + AIM), I evaluate each screen and each critical flow.

For each problem found, I document:

  • Location: Which screen/step it occurs
  • Description: What’s wrong
  • Heuristic violated: Which principle isn’t being met
  • Severity: 0 to 4 (Nielsen scale)
  • Screenshot: Visual evidence

5. Classify Severity

I use Nielsen’s classic scale:

Severity Description
0 Not a usability problem
1 Cosmetic problem - fix if time permits
2 Minor problem - low priority
3 Major problem - important to fix
4 Catastrophe - must fix before launch

6. Triangulate with Other Sources

I NEVER deliver a report based only on heuristics. I always cross-reference with:

  • Analytics data
  • User feedback (if it exists)
  • Competitor benchmark

7. Prioritize and Recommend

The final deliverable includes:

  • List of problems prioritized by severity
  • Evidence for each problem (screenshots, data)
  • Concrete improvement recommendations
  • Impact vs. effort matrix

Template and Tools

Basic Recording Template

You can use a simple spreadsheet with these columns:

ID Location Problem Description Heuristic Violated Severity (0-4) Evidence Recommendation
1 Checkout step 2 No progress indicator H1 - System status visibility 3 [screenshot] Add stepper with steps

Tools I Use

  • Screenshots: Any screenshot tool (Loom for videos)
  • Documentation: Google Sheets for recording, Notion for the report
  • Analytics: Google Analytics 4, Clarity (heatmaps and recordings)
  • Benchmark: Simply navigate reference sites and document best practices

Additional Resources

  • Heuristics posters: NNGroup has free posters you can have in your workspace
  • Hassan Montero’s guide: A detailed checklist in Spanish available at nosolousabilidad.com
  • Cugelman’s AIM Framework: Available on the AlterSpark and Behavioral Design Academy sites

Final Reflection

Heuristic evaluation is over 30 years old and remains relevant because it addresses something that doesn’t change: how we humans work.

Yes, ideally you’d have 3-5 evaluators. Yes, in consulting we often work solo. But that doesn’t mean the method doesn’t work. It means we have to be more strategic, more honest about limitations, and more rigorous in triangulating information.

At the end of the day, what matters is not checking off a checklist. It’s understanding the user and helping create products that truly work for them.

And that, with a bit of method and a lot of honesty, can be done even from a small consultancy in Chile.

References

  • Cugelman, B. (2013). Gamification: What It Is and Why It Matters to Digital Health Behavior Change Developers. JMIR Serious Games, 1(1), e3. https://doi.org/10.2196/games.3139
  • Hassan Montero, Y., & Martín Fernández, F. J. (2003). Guía de Evaluación Heurística de Sitios Web. No Solo Usabilidad, nº 2. https://www.nosolousabilidad.com/articulos/heuristica.htm
  • Nielsen, J. (1994). 10 Usability Heuristics for User Interface Design. Nielsen Norman Group. https://www.nngroup.com/articles/ten-usability-heuristics/
  • Nielsen, J., & Molich, R. (1990). Heuristic evaluation of user interfaces. CHI '90 Proceedings, 249-256.
  • Rubin, J., & Chisnell, D. (2008). Handbook of Usability Testing: How to Plan, Design, and Conduct Effective Tests (2nd ed.). Wiley.