Usability testing has been around for a while now, and there are some best practices to consider when selecting the right questions for your study. A well-conducted usability testing session is crucial for identifying user experiences, behavioral insights, and functionality issues, providing valuable feedback for design improvements and user experience enhancement.
Let’s start by looking at the most common question types to build a foundation; then, we’ll give you actual questions you can use in your usability tests.
Usability testing is a method used to evaluate the usability of a product, service, or digital experience by observing users completing specific tasks and answering questions while interacting with it. This process helps to understand how easy it is for the target audience to use the product or service, identify usability problems, and collect both qualitative and quantitative data to determine the usability of the design.
Usability testing is essential for ensuring that the product or service is user-friendly, meets the target audience's needs, and reduces development costs and time by identifying usability problems early on. By focusing on specific tasks, you can pinpoint exactly where users struggle and make necessary adjustments to improve the user experience.
Usability testing provides numerous benefits for user experience, including:
By incorporating usability testing into your development process, you can create a product that meets and exceeds user expectations, leading to higher satisfaction and engagement.
For getting feedback, there are two common question types. You’ll need to choose: open-ended vs closed-ended questions. While usability testing focuses on the effectiveness of a design, user testing assesses the necessity and utility of a product during early development stages.
Open-ended questions encourage free-form answers. They cannot be answered by “yes” or “no” responses. They start with words like “how,” “what,” “when,” “where,” “which,” and “who.”
TIP: Avoid “why” questions because they lead people to make up answers when they don't have a response. Instead, say something like, “please tell me more about that.”
When conducting qualitative usability research, you want to ask more open-ended questions because that's how you get human insight. Running small, moderated tests does not have statistical significance, so there's no point in getting answers that can be analyzed statistically. Instead, focus on digging deeper and getting richer data with open-ended questions.
Here are examples of open-ended questions:
Related reading: Ask the right questions in the right way
Closed-ended questions have a set of definitive responses, such as “yes,” “no,” “A,” “B,” “C,” etc. These questions are great for unmoderated surveys or type box responses because your users don't have to respond as much and instead offer validation or lack thereof. These answers can be analyzed statistically, so they're better suited for quantitative research than qualitative.
Here are examples of closed-ended questions:
When determining what types of questions to ask during usability testing, it’s important to consider a few different factors to ensure the prompts will gather the necessary insights. User testing questions play a crucial role in understanding how users interact with software and in selecting the right participants who resemble the target audience to provide relevant and actionable feedback.
Some factors to consider include:
“I start every project at the end – what do you need to solve for with this research? What decisions do you need to be able to make when it’s completed? It is critical that everyone involved is crystal clear on those things.” – Angie Amon, Mailchimp, UX Research Manager
To start, clarify the specific objectives of the usability testing. What do you hope to learn or achieve through the testing process? Aligning the questions with your team's research goals allows you to gather relevant and actionable insights.
Consider the characteristics of your target audience, such as:
Tailor the questions to address the intended user persona's needs, preferences, and behaviors.
Structure the questions around specific task scenarios you expect users to perform during the testing session. Focus on tasks representative of typical user interactions and goals to glean insights into the usability of key features and functionalities.
Use a variety of question types to gather both qualitative and quantitative data. Open-ended questions encourage users to provide detailed feedback and insights. Closed-ended questions such as Likert scale ratings—which have users mark answers based on a scale with responses like “strongly agree,” “agree,” “neutral,” “disagree,” and “strongly disagree”—allow more straightforward response quantification and comparison.
Additionally, write usability testing questions by addressing common inquiries and pitfalls in their formulation to enhance usability testing outcomes.
Consider incorporating standardized usability metrics, such as the System Usability Scale (SUS) or the Net Promoter Score (NPS), into your survey to assess overall usability and user satisfaction.
Organize the questions logically and sequentially to guide participants smoothly through the testing process. Start with broader, more general questions before delving into more specific usability aspects.
When conducting usability testing, consider potential biases influencing participants' responses. Avoid leading questions that may steer participants toward a particular response or influence their feedback. Instead, strive to ask neutral, unbiased questions that encourage honest and objective feedback from participants.
There are various usability test methods and approaches that can be used to evaluate the usability of a product or service. Each method has its own strengths and can be chosen based on the specific needs of your project.
When conducting usability testing, it’s essential to ask the right questions to gather valuable insights from users. This includes asking open-ended questions, avoiding leading questions, and using scales and multiple-choice questions to collect quantitative data. By using these methods and approaches, you can gather valuable user feedback and iterate on your design to create a better user experience.
Common usability testing questions take place in four phases:
Screener questions, or screeners, are questions intended to evaluate a contributor's qualifications and target specific groups of contributors. These multiple choice questions eliminate contributors who don't qualify to participate in your study.
Screeners allow you to find contributors based on their demographics or statistical data collected for a particular population by identifying variables and subgroups, like the following examples.
You can also filter contributors based on psychographics, data that collects and categorizes the population by using characteristics like interests, activities, and opinions:
To get started, identify the right target audience before creating screener questions, which ensures you get actionable feedback. For example, if most of your customers fall into a specific age range or geographic location, these might be the parameters you set. However, if you want to hear from those who may not be so familiar with your product for an unbiased outlook, think about enlisting users from outside the usual demographic.
The trick to effective screener questions is asking them in a way that identifies your audience without leading contributors to a particular answer or revealing specific information about your test.
For example, if you're looking to test out a new mobile app intended for parents who live in the midwestern United States, you want to find contributors who fit the criteria. Instead of asking someone if they live in the Midwest United States, you would ask in which regions you live and give answer choices for many different areas. Add distractor answers to your screener questions to deflect from the correct answer.
Here's how we would find our target audience of parents who live in the midwestern United States via screener questions:
2. Which of the following best describes your current status:
As you can see, we've woven the response we're looking for with distractor responses to increase our odds of getting the right contributors without revealing details about the test or who we're looking for.
Now that you've set up screener questions to find the ideal contributors, it's time to ask questions to learn more about your contributor before the test influences how they might answer. Pre-test usability questions give context to your contributor's actions and test answers. They can be open or closed-ended questions.
For example, you might want to know how experienced your contributor is with mobile apps before the usability study. This will help you better understand why they take specific actions.
Here are some examples of pre-test questions:
In-test usability questions are questions directly related to your testing objective. They should start general and get more specific. Always ask open-ended questions during your test.
Whether qualitative or quantitative, usability testing helps you understand the what, why, and how behind your customers and their actions. You can discover bugs or errors, get customer feedback, know your audience, learn whether something works as expected, and more.
When running an unmoderated usability test, you want to ensure your questions allow for open and honest feedback. Letting contributors know when you're open to critical or negative feedback is also a good idea. After a contributor finishes a task, here are some common open-ended questions to ask:
When running a moderated usability test, the moderator can probe deeper into the contributor's responses. A good rule of thumb is to stay quiet and let the contributor do most of the talking, but here are questions for promoting feedback:
Follow-up usability test questions are a set of questions that end the sessions. These might include clarifying or probing questions.
After a usability test, you have another chance to ask contributors about their experience for additional context. Get feedback on the experience overall and see if there's something they want to talk about that you didn't ask. Follow-up questions can be closed or open-ended.
As you can see, no test is complete without usability questions. For inspiration, take a look at some more usability testing examples. Or, to start testing, browse the UserTesting templates gallery for inspiration for your next project. The gallery contains pre-built templates designed by research experts, which can be used as-is or customized to fit your needs.
Get a free video of a real person reviewing your digital experience.