July 24, 2023
July 24, 2023
Usability testing is easy, they said. Just write some tasks, they said.
I oversimplified usability tests for a long time. While it is an easier methodology to learn and conduct, it isn't without challenges. It took me some time to perfect the craft of usability testing, and that included some very interesting sessions with participants, where I was just as confused as they were about tasks.
Additionally, since usability testing is such a popular methodology and one learned earlier, I defaulted to usability testing all the time, even when it wasn't relevant. This mismatch led to confusing results and broke my confidence in usability testing to the point where I avoided it.
Whether you want a refresher or are new to usability testing, this guide can help you learn or re-experience usability testing.
Psst…We’ve made a usability testing template that you can download here for free.
A stakeholder once asked me this question, and I was tongue-tied. I wanted to say, "Well, it's usability testing." Instead, I told them that usability testing was "testing the usability of a product."
And while that definition certainly is true, there is more to usability testing. For instance, what is usability? What are we testing, and how do we know what to test?
Usability is the ability for someone to:
For example, with a clothing company, a user would go to the website with the expectation of being able to purchase clothing. But just because a user can buy clothing, doesn't mean it's an easy or satisfactory experience.
So we can break usability down further into three areas:
Usability testing, whether you use metrics or run a qualitative usability test (more on that later), looks at these three factors to determine whether or not a product or service is usable.
So, what about those who browse e-commerce out of boredom (and no purchase intent)?
This leads to the next question...
As I mentioned above, there are many aspects you could test (even with a simple product!). But the point of a usability test is to ensure that users can complete their most common tasks to achieve their goals.
Again, one of the main goals for a clothing website would be to purchase clothing. You can have smaller goals within that larger goal, such as comparing clothing.
Then, you can break this down into the tasks people have to do to achieve those goals. For instance:
These are all tasks associated with the larger goal. With usability testing, we ask people to do these critical tasks to assess whether or not they can achieve them and the larger goal in an efficient, effective, and satisfactory way.
If someone can do these tasks, they can get to their expected outcome. However, if efficiency, effectiveness, or satisfaction suffer during this process, they may get frustrated and give up or go to a different website.
We've all encountered this—an infuriating user experience that made us rage-click, throw our phones (against a soft surface, of course), and give up on a product or service.
Now, when is usability testing an appropriate method? There are particular goals I always associate with usability testing, but before we dive in, let's understand what usability testing can help us with:
On the flip side, usability testing will NOT tell us:
Now, let's dive into some goals for usability testing because the best thing you can do is start with goals and a research plan.
These are the most common goals I use for usability tests:
The team at T-Mobile consistently needed robust insights on a tight timeline. See how they were able to shift from a reactive to a proactive research strategy, gather in-context mobile data quickly, and get their stakeholders more engaged with the data.
Once you've written your goals, it is time to think about recruitment and sample size.
Recruiting the right people is essential for a good user research study. Trust me, it's challenging to fill a 60-minute usability test with the wrong participants, plus it's awkward.
For usability tests, I ask myself the following questions, with some examples from above, to understand who my target users would be:
What are the particular behaviors I am looking for?
Have they needed to use the product? And in what timeframe?
What goals are important to our users?
What habits might they have?
You can create an effective screener survey when you know what you’re looking for.
One note about the sample size for usability testing: A general idea for evaluative research is testing with five people. When I review research plans, I often see five participants as the number for an evaluative study.
While this could be correct, it isn't a hard and fast rule. Like the above, if you pick five random people off the street to test your product, you likely won't find 85% of the issues.
The fine print behind five people per usability study is that it means five users per segment.
Here are the general guidelines for the top evaluative methods:
Task writing is the next step after recruitment and sorting out sample sizes. Task writing was my least favorite part for a while, because it’s more complicated than it seems on the surface.
Throughout the years, I have honed this skill through a lot of practice. If you’re looking to write great usability tasks, that would be my first piece of advice—practice, practice, practice.
These are the steps I go through when constructing my usability testing tasks:
Start with what you want the user's end goal to be, not the goal of the task. What does the user need (or want) to accomplish at the end of the action? What is their goal for using this product?
Instead of throwing participants into action with no relevant information, give them context on why they need to use the product. You can also consider the context and background information for why they would use the product in the real world.
Since you are recording metrics, you don't want to be vague in your instructions. If users need to input dates, locations, or particular data in a form, give them that information. You don't want the user to guess what you want them to do, resulting in skewed data.
If you’re trying to get someone to accomplish a task, make sure they can. There should be a reachable "end," which satisfies the participant and helps you record if the participants could complete the task.
Once you have brainstormed this information, write your task scenario. I have included some examples at the end of this article.
After writing down your task scenarios, perform a dry run with either internal employees, friends, family, or anyone that will give you their time. This dry run will allow you to practice the test flow, make sure the tasks make sense, and indicate whether there are too many.
*A common question is how many tasks should be in one usability test. It depends on the complexity of the tasks and how much time you have with the participant.
For a 45-60 minute session, I generally include five to seven tasks. By including a dry run in your process, you can know how many tasks you can fit into the session.
A potential task example for the above clothing company might look like this:
Winter is coming up, and you’re looking for a new winter coat to keep you warm that is under £150.
Usability tests can be either qualitative or more on the quantitative side if you incorporate usability metrics. However, I always recommend feeling comfortable with observational and qualitative usability testing before jumping in to track specific metrics. Also, if you are looking for qualitative feedback, usability metrics won't be a good fit.
When it comes to usability testing, we go back to our three cornerstones of usability:
Combining these metrics can help you highlight high-priority problem areas. For example, suppose participants respond confidently that they completed a task, yet most fail.
In that case, there is a vast discrepancy in how participants use the product, leading to problems. Let’s break up the metrics by area of usability testing:
Metrics won't be helpful if you’re looking for qualitative feedback and want to talk to participants about their experiences as they try tasks. You can't get qualitative feedback when you incorporate metrics into usability tests because you would be interrupting participants.
For instance, if you track task success or time on task, asking participants how they are feeling and what they are thinking during tasks can distract them and skew your data.
If you want numbers to assess the usability of your product, these metrics will be super helpful in bringing a quantitative twist to your user research. In addition, gathering this data and then testing again after improvements can also help toward benchmarking and proving the ROI of user research.
When using metrics, make sure your test evaluates effectiveness, efficiency, and satisfaction with data. If you want to add qualitative components, steer away from metrics or save some time for a qualitative portion after the metric-based usability test.
Once you conduct the usability test, it's time for the exciting process of unwinding your data into patterns and trends you can act on.
I usually focus on global tags and affinity diagramming when I synthesize qualitative usability testing. The most common global tags I use for usability testing are:
After each usability session, I do a quick debrief where I split the global tags into four quadrants and note what happened during that interview concerning each quadrant.
So, for example, if we were usability testing the clothing website, a session debrief (one participant) might look like this:
Completing this small synthesis session makes for easier work at the end of the study. Once you complete all the sessions and each debrief, bring all participants together to assess patterns and trends (the same thing three or more people are saying).
My go-to for reporting quantitative usability testing is a stoplight report.
A stoplight report:
The most valuable part of the stoplight approach is how visual it is. It can quickly provide a stakeholder with a holistic overview of how the usability test went.
A stoplight chart includes the following components (but you don't have to include all of them!):
Want to test it out? Use the template below for reporting your next quantitative usability study!
When you review basics and practice (as I said, practice is critical), you will naturally feel more comfortable with the different components of usability testing. Although it is a more straightforward method, it is still worth mastering—the right way.
Nikki Anderson-Stanier is the founder of User Research Academy and a qualitative researcher with 9 years in the field. She loves solving human problems and petting all the dogs.
To get even more UXR nuggets follow her on LinkedIn, or subscribe to her Substack.