Skip to main content
U.S. flag

An official website of the United States government

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Validate

Test a design hypothesis.

Card sorting

What

A categorization exercise in which participants divide concepts into different groups based on their understanding of those concepts.

Why

To gain insights from users about how to organize content in an intuitive way.

How to do it

There are two types of card sorting: open and closed. Most card sorts are performed with one user at a time, but you can also do the exercise with groups of two to three people.

Open card sort

  1. Give users a collection of content represented on cards.
  2. Ask users to separate the cards into whatever categories make sense to them.
  3. Ask users to label those categories.
  4. Ask users to tell you why they grouped the cards and labeled the categories as they did.

Closed card sort

  1. Give users a collection of content represented on cards.
  2. Ask users to separate the cards into a list of categories you have predefined.
  3. Ask users to tell you why they assigned cards to the categories they did.

Example from 18F

Additional resources

Government considerations

No PRA implications if done as directly moderated sessions. The PRA explicitly exempts direct observation and non-standardized conversation, 5 CFR 1320.3(h)3.

18F

Content highlighter testing

What

Users highlight text that they think is clear in one color. Then they highlight text they think is confusing in a different color.

Why

To understand which words and phrases are achieving their intended purpose. This test can help clarify how to improve the content.

How to do it

  1. Select content: Decide on the content you want to test.
  2. Select a tool: You can conduct this test in person, on paper, or online. If conducting the test online, choose an application that allows content highlighting and one that your users are familiar with.
  3. Define what you’re testing: Determine what qualities you want to test for. You can ask users to highlight clear content and then confusing content. Or, you can ask users to highlight content that makes them feel confident and then content that makes them feel less confident. It depends on what your goals are. We’ll use clear/confusing for this example.
  4. Pick the colors: Assign a color for each quality and explain what the colors mean. For example, “Use the green highlighter for information that is clear. Use the blue highlighter for information that is confusing.”
  5. Write instructions: If conducting the test online, prepare instructions that explain how the test will work, what the colors mean, how to highlight text, and how long the test will take.
  6. Start the test: Ask the user to read the text and highlight content according to your criteria.
  7. Discuss: You can ask users to discuss their highlights during the test. This can be useful for longer pieces of content. Or, you can wait until after the highlighting to discuss. Encourage users to describe why they highlighted passages. For example, “What did you find confusing about that sentence?”
  8. Analyze results: What did users think was clear? Confusing? How did the results differ across sessions? Identify priority content improvements based on the results.

Example from 18F

  • Project: pra.digital.gov
    • Problem: Site content was based on an internal document and not written for a public audience. The content needed to be approved by the general counsel before moving forward.
    • Highlighter test: The content was rewritten in plain language. Two attorneys reviewed it for clarity and accuracy using the highlighter method. Both attorneys reviewed the content at the same time, sitting at the same table, with their own copies of the content and colored highlighters. They discussed legal terms and wording with each other and the content strategist. In addition to improving the content, this test allowed us to avoid a later legal review.

Government considerations

In general, content highlighter testing doesn’t require PRA approval.

Read the PRA guidance on usability testing from the Office of Information and Regulatory Affairs (PDF). Consult the methods for recruiting and privacy for tips on getting input from the public.

18F

Multivariate testing

What

A test of variations to multiple sections or features of a page to see which combination of variants has the greatest effect. Different from an A/B test, which tests variation to just one section or feature.

Why

To incorporate different contexts, channels, or user types into addressing a user need. Situating a call to action, content section, or feature set differently can help you build a more effective whole solution from a set of partial solutions.

How to do it

  1. Identify the call to action, content section, or feature that needs to be improved to increase conversion rates or user engagement.
  2. Develop a list of possible issues that may be hurting conversion rates or engagement. Specify in advance what you are optimizing for (possibly through design hypothesis).
  3. Design several solutions that aim to address the issues listed. Each solution should attempt to address every issue by using a unique combination of variants so each solution can be compared fairly.
  4. Use a web analytics tool that supports multivariate testing, such as Google Website Optimizer or Visual Website Optimizer, to set up the testing environment. Conduct the test for long enough to produce statistically significant results.
  5. Analyze the testing results to determine which solution produced the best conversion or engagement rates. Review the other solutions, as well, to see if there is information worth examining in with future studies.

Government considerations

No PRA implications. The users are not being asked questions, so the PRA does not apply. See the methods for Recruiting and Privacy for more tips on taking input from the public.

18F

Success metrics

What

How to measure whether your design is successful in achieving the intended outcomes.

Why

No matter what you’re designing (software, a service, content, etc.), you need to define what your design should enable people to do and how you’ll measure whether it performs well.

How to do it

  1. Define the outcomes you expect when your design performs well. Invite stakeholders to brainstorm. Your outcomes may be based on your design hypotheses or user needs statements.
  2. List ways to measure whether your design is achieving expected outcomes. Consider quantitative and qualitative metrics:
    1. Quantitative metrics are numerical indicators (e.g. time on task, click-through rates).
    2. Qualitative metrics capture subjective feedback and insights (e.g. usability ratings, user satisfaction).
  3. Determine how you will collect data for each metric. This may involve user interviews, usability testing, surveys, and/or analytics. Choose metrics that are effective and measurable.
  4. Plan who will be responsible for data collection and how often metrics will be reviewed.
  5. Establish benchmarks for each metric. Benchmarks show whether the design meets peoples’ needs or should be further refined.
  6. Test your design, and evaluate success based on your measurement plan.
  7. Analyze the results against your benchmarks. Identify strengths and areas of improvement. Iterate on the design prototype. Refine metrics if necessary.

Additional resources

Government considerations

Surveys of more than nine people require PRA approval. Learn more at pra.digital.gov. Direct observation and non-standardized conversations like semi-structured interviews) are not subject to PRA.

Your prototype may not be the only effort being tested. People may be overwhelmed by the amount of data they’re asked to provide. Finding metrics that can be automatically collected can help relieve this burden.

18F

Usability testing

What

Observing users as they attempt to use a product or service while thinking out loud.

Why

To better understand how intuitive the team’s design is, and how adaptable it is to meeting user needs.

How to do it

  1. Pick what you’ll test. Choose something, such as a sketch, prototype, or even a "competitor’s product" that might help users accomplish their goals.
  2. Plan the test. Align your team on the scenarios the test will focus on, which users should participate (and how you’ll recruit them), and which team members will moderate and observe. Prepare a usability test script.
  3. Recruit users and inform their consent. Provide a way for potential participants to sign up for the test. Pass along to participants an agreement explaining what participation will entail. Clarify any logistical expectations, such as screen sharing, and how you’ll share links or files of whatever it is you’re testing.
  4. Run the tests. Moderators should verbally confirm with the participant that it’s okay to record the test, ask participants to think outloud, and guide the participant through the session. Observers should contribute to a rolling issues log and relay any in-session questions to the moderator, refraining from interrupting the session from the participant’s point of view. Engage your team in a post-interview debrief after each test.
  5. Discuss the results. Schedule a collaborative synthesis meeting to discuss issues you observed, and any questions these tests raise concerning user needs. Conclude the meeting by determining how the team will use what it learned in service of future design decisions. main

Examples from 18F

Additional resources

Government considerations

No PRA implications. First, any given usability test should involve nine or fewer users. Additionally, the PRA explicitly exempts direct observation and non-standardized conversation, 5 CFR 1320.3(h)3. It also specifically excludes tests of knowledge or aptitude, 5 CFR 1320.3(h)7, which is essentially what a usability test tests. See the methods for Recruiting and Privacy for more tips on taking input from the public.

18F

Visual preference testing

What

A method that allows potential users to review and provide feedback on a solution’s visual direction.

Why

To align the established branding guidelines and attributes of a solution with the way end users view the overall brand and emotional feel.

How to do it

  1. Create iterations of a style tiles or other asset that represent directions a final visual design may follow. If branding guidelines or attributes don’t exist, establish them with stakeholders beforehand.
  2. Interview participants about their reactions.
    • Ask questions as objectively as possible.
    • Align questions with the branding guidelines and attributes your project must incorporate.
    • As far as possible, allow participants to provide their feedback unmoderated or at the end of your research.
  3. Compare the results of your research with the agency’s published branding guidelines and attributes.
  4. Publish the results to the complete product team and decide which direction will guide future design efforts.

Government considerations

No PRA implications. The PRA explicitly exempts direct observation and non-standardized conversation, 5 CFR 1320.3(h)3. See the methods for Recruiting and Privacy for more tips on taking input from the public.

18F

18F Methods

An official website of the GSA’s Technology Transformation Services

Looking for U.S. government information and services?
Visit USA.gov