Your Usability Test Plan
Your Usability Test Plan
Views: 4,548
Table of Contents
While other types of testing are equally critical to product success, no one
likes a product with poor usability. The practice of usability has improved a
ton in the last few years, and a strong capability in this area is now standard
for any team in digital.
After reviewing this guide and the related practice, you will be able to:
1. Identify focused and appropriate usability testing for your product and
features
2. Design said testing around your user stories
3. Screen subjects and conduct usability testing
4. Interpret and iterate on the results of that testing
This accelerates learning on what interface patterns will work through early
user contact. Rather than having one or two shots at getting it right with
working software, you can get ten with the same amount of time and effort by
using prototypes.
You’ll start with an explicit prototype and test plan, but freely update them on
a test by test basis as you learn what’s working and your observations
encourage new ideas. Specific measurements are unimportant. Popular tools
are paper prototypes, PowerPoint/Keynotes prototypes, and various other
facile tools that discourage overemphasis on details and encourage
experimentation and variation. Exploratory testing drives to a decision about
the fundamental approach on an interface.
Assessment
As the name suggests, validation testing is for once you’re relatively sure
you’ve got an acceptable interface, and you want a) a final check on that and
b) some intuitive sense about the nature of what you’ll see with
measurements you take out in the wild (through Google Analytics, etc. with
users you don’t have the chance to meet and observe directly). Try hard to
leave yourself enough time to make at least a few minor changes before
releasing to the public- you’ll likely identify a few rough spots well worth
fixing.
You can also execute all three of these tests on a comparative basis
(comparison between alternatives).
How do I do it?
Preparation
User testing occurs relatively late in the overall discovery process. At this
point, you should have validated personas, problem scenarios, and value
propositions, something you can summarize in a Product Hypothesis. The
diagram below describes all this in terms of the convergence-divergence
pattern from design thinking. Your work on the Persona & Problem
Hypothesis should ultimately converge on a focal problem set, to which you
attach your value propositions. All this you can summarize in a Product
Hypothesis to connect problem with solution.
The Product Hypothesis will include a Value Hypothesis and that will have
certain assumptions attached to it. You should test those assumptions with
the quickest, least expensive MVP product proxies you think are viable to
provide a definitive, actionable result (see section on Value Hypothesis for
popular patterns). Once you’ve converged on which problems you can provide
adequately valuable solutions, encapsulate your solution ideas in agile user
stories. Then prototype, test, refine, and test some more!
Every substantial interaction you plan to test should have a user story
attached to it. The user stories will anchor your executions in the customer
narrative and validated learning you’ve developed. For every material
interface element, you should be able to answer these questions:
What problem are we solving for the user with this value
proposition?
Who is this user? Do we understand what makes them tick? What
shoes they might wear?
This will not only help you converge on more valuable solutions; it will also
help you zero in on the right place to start revisions when your execution
doesn’t go perfectly (and it never does!). Also, successful innovation is a loop,
not a line. I’ve presented most of the material as a sequence for clarity, but
the successful innovator is constantly looping back through the hypothesis
areas. People change, problems change (a little), and every solution is
temporary.
Execution
0. Prepare
User testing will deliver useful results on the investment just about all the
time. Don’t feel like you absolutely can’t do it if you haven’t completed the
items in the preceding section. That said, a little investment in those other
areas (personas, problem, MVP testing) will go a long way.
You’ve drafted one or more directions that are well supported by validated
learning and a review of comparables: Assessment
You think you’re done: Validation
This has several sections in addition to the test plan- I call use the term
‘research suite’ to designate the whole package for a given set of testing. At
first, it may look like a lot of stuff, but every time I’ve done user testing I’ve
found all this preparation well worthwhile (and the reverse is also true). The
example and template I reference are in the Venture Design
Template/Appendix B. Notes on preparing the various sections of the
research suite follow:
This is for you to use internally. Describe in the clearest possible terms what
you want to have happen as a result of the testing and your principal testing
focus (exploratory, assessment, or validation), linking that to your objectives.
Particularly if you’re early in a project, you may want to couple additional
customer discovery interviews (persona, problem hypothesis) with your user
testing. This is the place to explain that.
There are three general types of tests:- Exploratory: for learning about customer
problem scenarios in greater detail, sometimes with a paper or working prototype
This test suite is exploratory and we’re preceding the user testing with customer
discovery interviews from Appendix A [this is the appendix for customer discovery
interviews in the template] to deepen and align our view of personas and problem
scenarios with the exploratory test results.
:: Product Version
Note the version you’re planning to use for testing. See the next step (2B) on
defining a test release. Preview, though, so you don’t get stuck: this need not
be actual software or a real release- it could be paper prototypes or a
prototype you make in PowerPoint or Keynote.
:: Subjects
Define your target subject count in terms of the personas you’ve identified
(see above or tutorial on personas). You’re not trying to achieve statistical
significance in most cases (and almost always in the early phases), so 5-10
subjects is perfectly OK.
Since enabling the HR manager persona to be more effective is central to our value
proposition, our target weighting of subjects should reflects that. An ideal total and
mix of subjects would be:
Helen (or Hank!) the HR Manager 4Frank the Functional Manager 1-2The
screening question for both these subjects type is:
How many technical job candidates did you interview in the last month?
:: Research Composition
Here you will summarize everything that happens to a subject and how long
you think it will take. Here’s an example from Enable Quiz:
Duration
# Item Notes
(min.)
Here we will explain the objectives of the test and the
Intro. &
1 5 parameters of their participation. We’ll also obtain the
Explanation
designated release & consent form*.
Using the interview guide, we’ll spend a few minutes to
Discovery
2 20 discovery to improve our personas, problem scenarios
Questions
and user stories.
We’ll introduce the test scenario and then ask them to
3 Test Tasks 15
complete the Test Items.
Post-Test Make sure we ask if it is OK to follow-up with
4 5
Debrief additional questions.
:: Pre-Session Checklist
There’s nothing worse than starting of a test with something not being ready
or general not in the right state. This is a simple checklist to help you and/or
your collaborators make sure they don’t have any false starts. Here’s an
Example from Enable Quiz:
# Item Notes
Make sure have written versions of discovery and
1
test questions to refer to
– log in- make sure app is on
2 Make sure test instance is up and functional
starting page
Make sure recording equipment* is up and
3
functional
:: Session Design
This includes the intro you’ll do with subjects as well as the test items.
I strongly advise writing up the intro and practicing it- it takes work to put
yourself in the subject’s shoes and as things get busy (and repetitive) you’ll
easily miss things.
On the test items, you’ll notice each row has four items–
Research Objective: This will help keep you focused. Each item should have a
research objective (otherwise, why is it there?). If you’re running an
exploratory or assessment test, your user stories can provide a great anchor
for the objective (see example below).
Estimated vs. Actual Time: This is for setting expectations on duration as well
as evaluation. If you’re running an exploratory or assessment test, you’ll be
less concerned with actuals.
Notes: This is where you set up and design the testing. I like to break each of
these into a set of notes for the moderator and a set of target outputs. The
outputs should closely and obviously tie to the research objective.
Intro
Thanks for making time to take part in our study. My name’s [name] and this is
[observer]. [Explain participation and deal with consent form/obtain written
consent]*We’ll be using a test guide through the rest of this, so I hope you won’t
mind me referring to that.We’re here to learn about [an early version of a solution
that allows HR managers to assess the technical skill set of a job candidate through
an online quiz].I’m going to ask you some questions and give you some tasks. Feel
free to think out loud if you’re comfortable with that. We’re not here to test you and
there are no wrong answers. Our results just help us better understand our product
and its audience.The session will take roughly [40-60] minutes.Do you have any
questions before we start?Test Items
Est. v.
# Research Objective Actual Notes
(min.)
MODERATOR GUIDE
Let’s say your job is to create one of
these quizzes for an open position.
Here’s a description of the position
[Provide them sample job description
and let them review.]. Let me know
1 Exploratory Intro 5
when you’ve finished reviewing it and
if you have any questions.OUTPUT
Validation that the subject
understands their goal and the job
description, roughly as well as they
would in their current position.
MODERATOR GUIDE
Let’s say you want to create a new
Assess primary navigation for
2 2 quiz. What would you do?OUTPUT
new quiz creation
Assessment of primary navigation for
new quiz creation
MODERATOR GUIDE
How are we doing on this Tell me what you think you’re seeing
user story:As an HR here?Let’s say you wanted to choose
manager, I want to match an a set of quiz topics for the open
position you just reviewed. Show me
3 open position’s required 5 how you’d do that?OUTPUT
skills with quiz topics so I can An assessment of the user’s
create a quiz relevant for relationship to the available
candidate screening. affordances and their
appropriateness to the current user
? narratives and tasks.
… … … …
NOTES ON TAKEAWAYS
:: Post-Test Debrief
Do you really need this? Yes, probably. This is an after the fact checklist to
make sure you cover your bases: seeing if follow-up questions are OK,
compensating the subject (in whatever way you plan), seeing if they have
other thoughts, seeing if they have ideas on other subjects. Here’s an example
from Enable Quiz:
– Thanks so much. We’ll be using this to make the product and solution stuff like
documentation better.
– Would you mind if we send you follow-up questions?
– (if you’re giving them some kind of tangible thank you, make sure that gets done)
I don’t supply legal advice on this site and I don’t warrant these notes as fit
for legal compliance. As well it should be, recording individuals is subject to
various laws and regulations depending on who you are, who they are, where
you are, and how the recording will be used and stored (among other factors).
It’s important that you get advice from your legal counsel and maintain
transparency and applicable compliance with your subjects. At a minimum,
this means securing written releases for the recordings and making sure that
the recordings are stored and accessed securely (if you store them at all).
Regarding releases and consent, your specific compliance requirements will
vary, but here are a few sample consent forms from US institutions:
Usability.gov
Indiana University
If you’re working on an actual piece of software, test what’s current but don’t
make yourself crazy (and probably your subjects) by cutting it too close.
:: Paper Prototype
Yes, you can actually get meaningful test results from playing with pieces of
paper. It’s hard to believe until you do it.
To start, you need a prototype. The Balsamiq prototying process will serve
you well, assuming your subject is software. You will essentially prepare a set
of screens on paper and ask the user to interact with them- clicking (by
pointing) and typing (by writing with a pen).
Modularity and layering will serve you well in your preparation. I recommend
having a few templates that are regular paper layered on cardboard or a
similar substrate. This will make it easier to physically handle the prototypes
and exchange them with the user. The base template should look like your
target device- phone, tablet, laptop, etc.
Then layer basic screens on top of those (with light paste or spray adhesive
which you can buy at any craft store). On top of those you can layer additional
controls (Balsamiq lends itself to modular disposition of controls). And finally
on top of the controls you can layer Post-It’s (or strips thereof) onto which
users can type (photograph the results after tests and then just replace the
Post-It’s).
Try it a few times and you’ll probably find you’re not uncomfortable with the
process.
This is the same basic idea as paper prototyping but you’re simulating the
interaction with inter-slide links on PPT or Keynote. The advantage is that
everything’s on the computer if you’re not a glue-and-scisssors fan, and the
experience may feel more real to subjects. The disadvantage is that the
linking can get confusing, improvisation is harder, and if you want the user to
fill out text you’ll need to have a paper form for them anyway.
Create the various (static) screens you want as slides within your application
of choice. Then add inter-slide links. In the current version of PowerPoint
(Mac; I’ll guess it’s the same on PC but haven’t been able to check), you do
this by:
– selecting ‘Hyperlink’
– and using the ‘Locate’ button to find an ‘Anchor’ (you’ll need to click the
right triangle to unfurl the list of slides).
Few of you will have access to observation booths etc., so I’ll skip that.
If you have the team size, separating the facilitator and observer/note-taker
functions is very helpful, leaving the facilitator free to focus on the experience
of the subject.
Make sure the facilitator is close by, but ideally not immediately visible or
over the shoulder of the subject. A good location is between the subjects 4/5-
o’clock and 2′-4′ distant. The observers will generally sit behind the subject-
as far away as possible where they can still see what’s happening.
A simple PC/Mac with a web-cam will do fine. For recording screen activity
and a web-cam feed on a PC, I like Camtasia Studio. For the Mac, I use
ScreenFlow. Make sure you have everything recording and rendering the way
you expect beforehand.
Note: You have serious obligations (ethical and legal) to steward and
safeguard your subjects’ privacy and obtain their explicit agreement on
participation, particularly if you’re recording. See the above note on
‘Recording & Compliance’.
4. Obtain Subjects
First off, if you’re trying out a new test set up (not to mention if you’ve never
done this before), find some subjects where you can ‘test the test’. This is
anyone who could plausibly use the product, even if they don’t well represent
one of your target personas. Things will break, you’ll fix them, don’t worry,
it’s natural.
If you have a research suite along the lines of what’s above/in the template,
then you have a plan.
In working the plan, practice is the best tutor. Be careful not to coach subjects
too much or make them feel judged. It’s painful to watch them struggle with
something they don’t understand, but better to learn about that now than
subject every future user of your product to it! Give them time to work
through confusion. Eventually (set a threshold for yourself) you’ll need to
help them move forward, but make sure you don’t do it too soon.
Don’t forget to thank your subject, compensate them (in whatever way you
plan), and ask them if follow-up’s are OK.
I recommend doing this right away. Most of the important insights you’ll
have, you’ll have on the spot.
Validation (Invalidation)
As with any test, conclusions are the point. Success/a good result will vary by
test type.
Exploratory: The results should help you better understand the likely journey
of a typical user and, depending on where you are in designing/prototyping
the interface, whether you’re headed in a workable direction. Comparison
tests here are highly desirable given their low cost and possible impact.
Validation: Here you’ll generally have a quantitative target for time spent per
task and in total on your major experience arcs. Validation is being within a
reasonable deviation from that.
Pro Tips
It is cumbersome, but make sure you’re actually getting the subjects to engage
with all the individual interface components, prompting them (without
leading them) through your test plan. While you’re validating a rough general
concept, the way you do that is through careful test design and observation of
what the subject actually does (vs. says).
Supply Motivation
You’re testing usability, not motivation and this is not the right way to test
motivation (for that, see Lean Startup and the use of MVP’s). Make sure you
give the user an intro that tells them the specific situation they’re in and gives
them their goal. Then with a minimum of prompting see if they can realize
those goals with your current interface. Don’t ask them if they want to do x or
y and don’t ask them to think up their own inputs and parameters unless you
have a really good reason.
If you have a big red button, don’t prompt your user with ‘Would you show
me how you would press the big red button?’. You already know what’s going
to happen. Instead, prompt your user with what you’re telling them is their
goal and find alternative but relevant language to describe your interface
instead of keying to specific labels on the page.
This test suite is exploratory and we’re preceding the user testing with
customer discovery interviews to deepen and align our view of personas and
problem scenarios with the exploratory test results.
Product Version
We’ll be using version [0.1] of the product for this exploratory test. [NOTE:
They could easily be using paper or PowerPoint prototypes as this stage as
well]
Subjects
Since enabling the HR manager persona to be more effective is central to our
value proposition, our target weighting of subjects should reflects that. An
ideal total and mix of subjects would be:
Research Composition
Duration
# Item Notes
(min.)
Here we will explain the objectives of the test and
Intro. &
1 5 the parameters of their participation. We’ll also
Explanation
obtain the designated release & consent form*.
Using the interview guide, we’ll spend a few minutes
Discovery
2 20 to discovery to improve our personas, problem
Questions
scenarios and user stories.
We’ll introduce the test scenario and then ask them
3 Test Tasks 15
to complete the Test Items.
Post-Test Make sure we ask if it is OK to follow-up with
4 5
Debrief additional questions.
Pre-Session Checklist
# Item Notes
1 Make sure have written versions of discovery and
test questions to refer to
– log in
2 Make sure test instance is up and functional
– make sure app is on
starting page
Make sure recording equipment* is up and
3
functional
Session Design
Intro
Thanks for making time to take part in our study. My name’s [name] and this
is [observer]. [Explain participation and deal with consent form/obtain
written consent]*
We’ll be using a test guide through the rest of this, so I hope you won’t mind
me referring to that.
We’re here to learn about [an early version of a solution that allows HR
managers to assess the technical skill set of a job candidate through an online
quiz].
I’m going to ask you some questions and give you some tasks. Feel free to
think out loud if you’re comfortable with that. We’re not here to test you and
there are no wrong answers. Our results just help us better understand our
product and its audience.
Test Items
Est. v.
Actual
# Research Objective Notes
(min.)
MODERATOR GUIDE
NOTES ON TAKEAWAYS
Post-Test Debrief
– Thanks so much. We’ll be using this to make the product and solution stuff
like documentation better.
– (if you’re giving them some kind of tangible thank you, make sure that gets
done)
* I don’t supply legal advice on this site and I don’t warrant these notes as fit
for legal compliance. As well it should be, recording individuals is subject to
various laws and regulations depending on who you are, who they are, where
you are, and how the recording will be used and stored (among other factors).
It’s important that you get advice from your legal counsel and maintain
transparency and applicable compliance with your subjects. At a minimum,
this means securing written releases for the recordings and making sure that
the recordings are stored and accessed securely (if you store them at all).
Regarding releases and consent, your specific compliance requirements will
vary, but here are a few sample consent forms from US institutions:
Usability.gov
Indiana University
Screener
How many times last month did you post to social media for your business?
What services did you use?
Product Version
We’ll be using version [x.y] of the product for this exploratory test. [NOTE:
They could easily be using paper or PowerPoint prototypes as this stage as
well]
Subjects
Our core subject has a small company or personal brand they’re promoting.
[XYZ], is central to our value proposition of [ABC] so we’re targeting a
composition of subjects as follows (organized against our personas):
Research Composition
Duration
# Item Notes
(min.)
Here we will explain the objectives of the test and
Intro. &
1 5 the parameters of their participation. We’ll also
Explanation
obtain the designated release & consent form*.
We’ll introduce the test scenario and then ask them
2 Test Tasks 15
to complete the Test Items.
Post-Test Make sure we ask if it is OK to follow-up with
4 5
Debrief additional questions.
Pre-Session Checklist
# Item Notes
Make sure have written versions of discovery and test
1
questions to refer to
– log in
Session Design
Intro
Thanks for making time to take part in our study. My name’s [name].
[Explain participation and deal with consent form/obtain written consent]*
We’ll be using a test guide through the rest of this, so I hope you won’t mind
me referring to that.
We’re here to learn about a product that helps individuals and teams manage
social media accounts.
I’m going to ask you some questions and give you some tasks. Feel free to
think out loud if you’re comfortable with that. We’re not here to test you and
there are no wrong answers. Our results just help us better understand our
product and its audience.
Test Items
Est. v.
# Research Objective Actual Notes
(min.)
MODERATOR GUIDE
What do you think you’re seeing
How are we doing on this user here?Let’s say you want to sign
story: up. Would you show me how
you’d do that?
As a Sam the Small Business
1 n/a OUTPUT
Owner, I want to sign up for the
User’s understanding of the
service, so I can give it a try.’
landing page. Validation that the
? subject can complete and
understands the initial signup
process.
How are we doing on this user MODERATOR GUIDE
story: Let’s say you wanted to add your
Twitter and LinkedIn accounts.
As a Sam the Small Business Would you show me how you
2 Owner, I want to connect my n/a would you do that?
social media accounts, so I can OUTPUT
create automated posts to them.’ An assessment of whether the
user can connect accounts and
? understands the process.
MODERATOR GUIDE
Let’s say you wanted to compose
and schedule a posting for 2 days
How are we doing on this user now at 8AM [local time zone].
story: The text is:‘Free fries from 3-6pm
today!’and the url is
As Sam the Small Business www.alexandercowan.com
Owner, I want to schedule some I’ll help you make sure it’s not
4 posts for two days from now at n/a accidentally posted and we’ll
8AM [local time zone], so I know delete it at the end. Let’s say you
it’s going to post at the time I want it to post to both Twitter
want. and LinkedIn. Would you show
me how you’d do that?
? OUTPUT
An assessment of whether the
user can schedule posts and
understands the process.
MODERATOR GUIDE
How are we doing on this user
Let’s say you wanted to change
story:
the items you just scheduled to
As Sam the Small Business post at 10am instead of 8am.
Owner, I want to update the time Would you show me how you
5 on a scheduled event, so I know n/a would do that?
it’s going to post at the time I
want. OUTPUT
An assessment of whether the
? user can connect find scheduled
posts and update their post
times.
MODERATOR GUIDE
How are we doing on this user Let’s say you wanted to remove
story: the LinkedIn post so nothing
goes out at all. Would you show
As Sam the Small Business me how you would do that?
6 Owner, I want to remove a n/a (Repeat for Twitter)
scheduled post, so I know it’s not
going to post. OUTPUT
An assessment of whether the
? user can locate and delete posts
accounts and understands the
process.
NOTES ON TAKEAWAYS
Post-Test Debrief
– Thanks so much. We’ll be using this to make the product and solution stuff
like documentation better.
– Would you mind if we send you follow-up questions?
– (if you’re giving them some kind of tangible thank you, make sure that gets
done)
– (make sure to delete their accounts and all login, password, and personal
identifying data)
* I don’t supply legal advice on this site and I don’t warrant these notes as fit
for legal compliance. As well it should be, recording individuals is subject to
various laws and regulations depending on who you are, who they are, where
you are, and how the recording will be used and stored (among other factors).
It’s important that you get advice from your legal counsel and maintain
transparency and applicable compliance with your subjects. At a minimum,
this means securing written releases for the recordings and making sure that
the recordings are stored and accessed securely (if you store them at all).
Regarding releases and consent, your specific compliance requirements will
vary, but here are a few sample consent forms from US institutions:
Usability.gov
Indiana University