March 20, 2023
10

Using User Engagement Metrics to Improve Your Website's User Experience

Are your users engaged in your website? The success of your website will largely depend on your answer. After all, engaged users are valuable users; they keep coming back and will recommend your site to colleagues, friends, and family. So, if you’re not sure if your users are engaged or not, consider looking into your user engagement metrics.

User engagement can be measured using a number of key metrics provided by website analytics platforms. Metrics such as bounce rate, time on page, and click-through rate all provide clues to user engagement and therefore overall website user experience.

This article will help you understand user engagement and why it’s important to measure. We’ll also discuss how to apply user engagement insights to improve website success. Combining a little bit of data with some user research is a powerful thing, so let’s get into it.

Understanding User Engagement Metrics 📐

User engagement metrics provide valuable insight for both new and existing websites. They should be checked regularly as a sort of ‘pulse check’ for website user experience and performance. So, what metrics should you be looking at? Website metrics can be overwhelming; there are hundreds if not thousands to analyze, so let’s focus on three:

Bounce rate


Measures the percentage of users that visit just one page on your site before leaving. If your bounce rate is high it suggests that users aren’t finding the content relevant, engaging, or useful. It points to a poor initial reaction to your site and means that users are arriving, making a judgment about your design or content, and then leaving.

Time on page


Calculated by the time difference between the point when a person lands on the page and when they move on to the next one. It indicates how engaging or relevant individual pages on your website are. Low time on page figures suggest that users aren’t getting what they need from a certain page, either in terms of the content, the aesthetics, or both.

Click-through rate


Click-through rate compares the number of times someone clicks on your content, to the number of impressions you get (how many times an internal link or ad was viewed). The higher the rate, the better the engagement and performance of that element. User experience design can influence click-through rates through copywriting, button contrasts, heading structure, navigation, etc.

Conversion rate


Conversion rates are perhaps the pinnacle of user engagement metrics. Conversion rate is the percentage of users that perform specific tasks you define. They are therefore dictated by your goals, which could include form submissions, transactions, etc. If your website has high conversion rates, you can be fairly confident that your website is matching your users’ needs, requirements, and expectations.

But how do these metrics help? Well, they don’t give you an answer directly. The metrics point to potential issues with website user experience. They guide further research and subsequent updates that lead to website improvement. In the next section, we’ll discuss how these and others can support better website user experiences.

Identifying Areas for Improvement 💡

So, you’ve looked at your website’s user engagement metrics and discovered some good, and some bad. The good news is, there’s value in discovering both! The catch? You just need to find it. Remember, the metrics on their own don’t give you answers; they provide you direction.

The ‘clues’ that user engagement metrics provide are the starting point for further research. Remember, we want to make data-driven decisions. We want to avoid making assumptions and jumping to conclusions about why our website is reporting certain metrics. Fortunately, there are a bunch of different ways to do this.

User research data can be gathered by using both qualitative and quantitative research techniques. Insights into user behavior and needs can reveal why your website might be performing in certain ways. Research can include both qualitative and quantitative techniques.

Qualitative research techniques

  • Usability test – Test a product with people by observing them as they attempt to complete various tasks.
  • User interview – Sit down with a user to learn more about their background, motivations and pain points.
  • Contextual inquiry – Learn more about your users in their own environment by asking them questions before moving onto an observation activity.
  • Focus group – Gather 6 to 10 people for a forum-like session to get feedback on a product.

Quantitate research techniques

  • Card sorts – Find out how people categorize and sort information on your website.
  • First-click tests – See where people click first when tasked with completing an action.
  • A/B tests – Compare 2 versions of a design in order to work out which is more effective.
  • Clickstream analysis – Analyze aggregate data about website visits.
  • Tree-testing - Test your site structure using text-only categorization and labels

The type of research depends on what question you want to answer. Being specific about your question will help you identify what research technique(s) to deploy and ultimately the quality of your answer. If you’re serious about website improvement; identify problem areas with user engagement metrics, and investigate how to fix them with user research.

Optimizing Content and Design

If you have conducted user research and found weak areas on your website, there are many things to consider. Three good places to start are navigation, content, and website layout. Combined, these have a huge impact on user experience and can be leveraged to address disappointing engagement metrics.

Navigation


Navigation is a crucial aspect of creating a good user experience since it fundamentally connects pages and content which allows users to find what they need. Navigation should be simple and easy to follow, with important information/actions at the top of menus. Observing the results of card sorting, tree testing, and user testing can be particularly useful in website optimization efforts. You may find that search bars, breadcrumb trails, and internal links can also help overcome navigation issues.

Content


Are users seeing compelling or relevant content when they arrive on your site? Is your content organized in a way that encourages further exploration? Card sorting and content audits are useful in answering these questions and can help provide you with the insights required to optimize your content. You should identify what content might be redundant, out of date, or repetitive, as well as any gaps that may need filling.

Layout


A well-designed layout can improve the overall usability of a website, making it easier for users to find what they're looking for, understand the content, and engage with it. Consider how consistent your heading structures are and be sure to use consistent styling throughout the site, such as similar font sizes and colors. Don’t be afraid to use white space; it’s great at breaking up sections and making content more readable.

An additional factor related to layout is mobile optimization. Mobile-first design is necessary for apps, but it should also factor into your website design. How responsive is your website? How easy is it to navigate on mobile? Is your font size appropriate? You might find that poor mobile experience is negatively impacting user engagement metrics.

Measuring Success 🔎

User experience design is an iterative, ongoing process, so it’s important to keep a record of your website’s user experience metrics at various points of development. Fortunately, website analytics platforms will provide you with historic user data and key metrics; but be sure to keep a separate record of what improvements you make along the way. This will help you pinpoint what changes impacted different metrics.

Define your goals and create a website optimization checklist that monitors key metrics on your site. For example, whenever you make an update, ensure bounce rates don’t exceed a certain number during the days following; check that your conversion rates are performing as they should be; check your time on sites hasn’t dropped. Be sure to compare metrics between desktop and mobile too.

User’s needs and expectations change over time, so keep an eye on how new content is performing. For example, which new blog posts have attracted the most attention? What pages or topics have had the most page views compared to the previous period? Tracking such changes can help to inform what your users are currently engaged in, and will help guide your user experience improvements.

Conclusion 🤗

User engagement metrics allow you to put clear parameters around user experience. They allow you to measure where your website is performing well, and where your website might need improving. Their main strength is in how accessible they are; you can access key metrics on website analytics platforms in moments. However, user engagement metrics on their own may not reveal how and why certain website improvements should be made. In order to understand what’s going on, you often need to dig a little deeper.

Time on page, bounce rate, click-through rate, and conversion rates are all great starting points to understand your next steps toward website improvement. Use them to define where further research may be needed. Not sure why your average pages per session is two? Try conducting first-click testing; where are they heading that seems to be a dead end? Is your bounce rate too high? Conduct a content audit to find out if your information is still relevant, or look into navigation roadblocks. Whatever the question; keep searching for the answer.

User engagement metrics will keep you on your toes, but that’s a good thing. They empower you to make ongoing website improvements and ensure that users are at the heart of your website design. 

Share this article
Author
Optimal
Workshop

Related articles

View all blog articles
Learn more
1 min read

5 tips for running an effective usability test

Usability testing is one of the best ways to measure how easy and intuitive to use something is by testing it with real people. You can read about the basics of usability testing here.

Earlier this year, a small team within Optimal Workshop completely redesigned the company blog. More than anything, we wanted to create something that was user-friendly for our readers and would give them a reason to return. I was part of that team, and we ran numerous sessions interviewing regular readers as well as people unfamiliar with our blog. We also ran card sorts, tree tests and other studies to find out all we could about how people search for UX content. Unsurprisingly, one of the most valuable activities we did was usability testing – sitting down with representative users and watching them as they worked through a series of tasks we provided. We asked general questions like “Where would you go to find information about card sorting”, and we also observed them as they searched through our website for learning content.

By stripping away any barriers between ourselves and our users and observing them as they navigated through our website and learning resources, as well as those of other companies, we were able to build a blog with these people’s behaviors and motivations in mind.

Usability testing is an invaluable research method, and every user researcher should be able to run sessions effectively. Here are 5 tips for doing so, in no particular order.

1. Clarify your goals with stakeholders

Never go into a usability test blind. Before you ever sit down with a participant, make sure you know exactly what you want to get out of the session by writing down your research goals. This will help to keep you focused, essentially giving you a guiding light that you can refer back as you go about the various logistical tasks of your research. But you also need to take this a step further. It’s important to make sure that the people who will utilize the results of your research – your stakeholders – have an opportunity to give you their input on the goals as early as possible.

If you’re running usability tests with the aim of creating marketing personas, for example, meet with your organization’s marketing team and figure out the types of information they need to create these personas. In some cases, it’s also helpful to clarify how you plan to gather this data, which can involve explaining some of the techniques you’re going to use.

Lastly, find out how your stakeholders plan to use your findings. If there are a lot of objectives, organize your usability test so you ask the most important questions first. That way, if you end up going off track or you run out of time you’ll have already gathered the most important data for your stakeholders.

2. Be flexible with your questions

A list of pre-prepared questions will help significantly when it comes time to sit down and run your usability testing sessions. But while a list is essential, sometimes it can also pay to ‘follow your nose’ and steer the conversation in a (potentially) more fruitful direction.

How many times have you been having a conversation with a friend over a drink or dinner, only for you both to completely lose track of time and find yourselves discussing something completely unrelated? While it’s not good practice to let your usability testing sessions get off track to this extent, you can surface some very interesting insights by paying close attention to a user’s behavior and answers during a testing session and following interesting leads.

Ideally, and with enough practice, you’ll be able to answer your core (prepared) questions and ask a number of other questions that spring to mind during the session. This is a skill that takes time to master, however.

3. Write a script for your sessions

While a usability test script may sound like a fancy name for your research questions, it’s actually a document that’s much more comprehensive. If you prepare it correctly (we’ll explain how below), you’ll have a document that you can use to capture in-depth insights from your participants.

Here are some of the key things to keep in mind when putting together your script:

  • Write a friendly introduction – It may sound obvious, but taking the time to come up with a friendly, warm introduction will get your sessions off to a much better start. The bonus of writing it down is that you’re far less likely to forget it!
  • Ask to record the session – It’s important to record your session (whether through video or audio), as you’ll want to go back later and analyze any details you may have missed. This means asking for explicit permission to record participants. In addition to making them feel more comfortable, it’s just good practice to do so.
  • Allocate time for the basics – Don’t dive into the complex questions first, use the first few minutes to gather basic data. This could be things like where they work and their familiarity with your organization and/or product.
  • Encourage them to explain their thought process – “I’d like you to explain what you’re doing as you make your way through the task”. This simple request will give you an opportunity to ask follow-up questions that you otherwise may not have thought to ask.
  • Let participants know that they’re not being tested – Whenever a participant steps into the room for a test, they’re naturally going to feel like they’re being tested. Explain that you’re testing the product, not them. It’s also helpful to let them know that there are no right or wrong answers. This is an important step if you want to keep them relaxed.

It’s often easiest to have a document with your script printed out and ready to go for each usability test.

4. Take advantage of software

You’d never see a builder without a toolbox full of a useful assortment of tools. Likewise, software can make the life of a user research that much easier. The paper-based ways of recording information are still perfectly valid, but introducing custom tools can make both the logistics of user research and the actual sessions themselves much easier to manage.

Take a tool like Calendly, for example. This is a powerful piece of scheduling software that almost completely takes over the endless back and forth of scheduling usability tests. Calendly acts as a middle man between you and your participants, allowing you to set the times you’re free to host usability tests, and then allowing participants to choose a session that suits them from these times.

Our very own Reframer makes the task of running usability tests and analyzing insights that much easier. During your sessions, you can use Reframer to take comprehensive notes and apply tags like “positive” or “struggled” to different observations. Then, after you’ve concluded your tests, Reframer’s analysis function will help you understand wider themes that are present across your participants.

There’s another benefit to using a tool like Reframer. Keeping all of your notes in place will mean you easily pull up data from past research sessions whenever you need to.

5. Involve others

Usability tests (and user interviews, for that matter) are a great opportunity to open up research to your wider organization. Whether it’s stakeholders, other members of your immediate team or even members of entirely different departments, giving them the chance to sit down with users will show them how their products are really being used. If nothing else, these sessions will help those within your organization build empathy with the people they’re building products for.

There are quite a few ways to bring others in, such as:

  • To help you set up the research – This can be a helpful exercise for both you (the researcher) and the people you’re bringing in. Collaborate on the overarching research objectives, ask them what types of results they’d like to see and what sort of tasks they think could be used to gather these results.
  • As notetakers – Having a dedicated notetaker will make your life as a researcher significantly easier. This means you’ll have someone to record any interesting observations while you focus on running the session. Just let them know what types of notes you’d like to see.
  • To help you analyze the data – Once you’ve wrapped up your usability testing sessions, bring others in to help analyze the findings. There’s a good chance that an outside perspective will catch something you may miss. Also, if you’re bringing stakeholders into the analysis stage, they'll get a clearer picture of what it means and where the data came from.

There are myriad other tips and best practices to keep in mind when usability testing, many of which we cover in our introductory page. Important considerations include taking good quality notes, carefully managing participants during the session (not giving them too much guidance) and remaining neutral throughout when answering their questions. If you feel like we’ve missed any really important points, feel free to leave a comment!

Read more

Learn more
1 min read

A comprehensive look at usability testing

Usability testing has an important role in UX and if you’re new to it, this article gives you a solid introduction to it with practical tips, a checklist for success and a look at our remote testing tool, Treejack.

Concepts of usability testing

Usability testing is the process of evaluating a product or service with users prior to implementation. The goal of usability testing is to identify any usability issues before the product or service is released into the world for use. Usability testing is a research activity that results in both quantitative and qualitative insights and can be used to gauge user satisfaction.A typical usability testing session is moderated and involves a participant, a facilitator and an observer. The facilitator leads the session and the observer takes notes while the participant completes the task based scenario.

While this is common, usability testing is scalable and the possible approaches are endless giving you the flexibility to work with the resources you have available—sometimes one person performs the role of facilitator and observer!Location also varies for usability testing. For example, you might conduct your testing in a lab environment or you might talk to users in a specific environment. It also worth noting that not all usability testing sessions are moderated—more about this later.Usability testing usually occurs multiple times during the design process and can be conducted anytime you have a design you would like to test.

User research activities like a focus group for example, are conducted early in the design process to explore and gain understanding before ideas are generated. Usability testing is about testing those ideas as early and as often as possible. From a fully functioning digital prototype to a simple hand drawn wireframe on paper, nothing is too unrefined or too rough to be tested.

Developing a usability test plan

Before you start a round of usability testing, you need to develop a usability test plan. The usability test plan will keep you organised and is an opportunity to define roles and set clear expectations upfront. The first step in developing this is to hold a meeting with your team and stakeholders to discuss what you are going to do and how you plan to achieve it. Following this meeting, a document outlining the usability test plan as it was discussed is created and shared with the group for review. Any changes suggested by the group are then added to the final document for approval from the relevant stakeholders.

What to include in your usability test plan:

  • The goal, scope and intent of the usability testing
  • Constraints impacting upon testing
  • Details on what will be tested eg wireframes
  • Schedule and testing location
  • Associated costs eg participant recruitment
  • Facilitator and observer details for each session
  • Session details
  • Participant recruitment approach
  • Equipment
  • Details of any documentation to be produced eg a report

Usability testing questions

Once you have developed your test plan, you need to create a list of questions and task based scenarios for the testing session. These form the structure for your testing and provide a framework of consistency across all testing sessions in the study.The questions serve as a warm up to ease the participant into the session and can also provide insights on the user that you may not have had before. These questions can be a combination of open and closed questions and are especially useful if you are also developing personas for example. Some examples of what you might ask include:

  • Tell me about a recent experience you had with this product/service
  • Do you currently use this product/service?
  • Do you own a tablet device?

The purpose of the task based scenarios is to simulate a real life experience as closely as possible. They provide a contextual setting for the participant to frame their approach and they need to be realistic—your participant needs an actionable starting point to work from. A good starting point for task based scenario development would be to look at a use case.It is also important that you avoid using language that provides clues to the solution or leads your participant as this can produce inaccurate results. An example of a task based scenario would be:You’re planning a Christmas vacation to New Zealand for your family of two adults and 4 children. Find the lowest priced airfares for your trip.

Usability testing software: Tree testing

Treejack is a remote information architecture (IA) validation tool that shows you exactly where users are getting lost in your content. Knowing this will enable you to design a structure for your website that makes sense to users before moving on to the user interface (UI) design.Treejack works like a card sort in reverse. Imagine you have just completed a card sort with your users to determine your IA and you are now working backwards to test that thinking against real world task based scenarios. Treejack does this using a text-based version of your IA that is free from distracting visual aids such as navigation and colour allowing you to determine if your structure is usable from the ground up. A Treejack study is structured around task based scenarios and comes with the option to include pre and post study questionnaires.

treejack task image
Usability testing with Treejack

As a remote testing tool, Treejack is unmoderated and provides the opportunity to reach a much larger audience because all you have to do is share a link to the study with your participants to gain insights. You also have the option of handing the task of targeted participant recruitment over to Optimal Workshop.Once launched and shared with participants, Treejack takes care of itself by recording the results as they come in giving you the freedom to multitask while you wait for the testing to finish.

The results produced by Treejack are not only detailed and comprehensive but are also quite beautiful. The story of your participants’ journey through your testing activity is told through pietrees. A pietree is a detailed pathway map that shows where your participants went at each fork in the road and their destinations. They allow you to pinpoint exactly where the issues lie and are a powerful way to communicate the results to your team and stakeholders.

bananacom pie tree
Treejack presents your results using pietrees

Treejack also provides insights into where your participants landed their first click and records detailed information on pathways followed by each individual participant.

bananacom paths
Treejack records full details of the paths followed by every participant

Usability testing checklist

The following checklist will help ensure your usability testing process runs smoothly:

  • Meet with team and stakeholders
  • Determine goals, scope and intent of usability testing
  • Decide how many sessions will be conducted
  • Create usability testing schedule
  • Select facilitators and observers for each session if applicable
  • Develop and complete a usability test plan
  • Determine test questions and scenarios
  • Recruit participants for testing
  • Gather equipment required for testing if applicable
  • Book testing location if applicable
  • Keep a list of useful contact details close by in case you need to contact anyone during testing
  • Complete a dry run of a testing session with a team member to ensure everything works before actual testing begins
  • Organise debrief meetings with observers after each testing session
  • Set aside time to analyse the findings
  • Document and present findings to team and relevant stakeholders

Related reading

Learn more
1 min read

Content design for startups: how to work lean, have maximum impact, and get all the high-fives

When you have a small design team or none at all, how do you ensure that your content is consistent, has the right tone, and is captivating? It can be difficult, but it doesn’t have to be! Julia Steffen, Principal Content Designer at Varis, spoke at UX New Zealand, the leading UX and IA conference in New Zealand hosted by Optimal Workshop, about how startups can achieve impactful content and delight users. 

In her talk, Julia shares her most useful tips, tricks, and rules of thumb to ensure meaningful content design. She also shares some helpful tools to achieve maximum efficiency.

Julia Steffen bio 🎤

Julia has worked in content for 10+ years at St.Jude, Wunderman Thompson, MetaLab, and Grubhub. She is based in the United States and is the Principal Content Designer at Varis.

Contact Details:

Email address: julia.steffen@govaris.com

You can find Julia on LinkedIn

Content design for startups - How to work lean, have maximum impact, and get all the high-fives ✋🏽✋🏻✋🏿

Why should you care about content design? Julia argues that “content design is product success”. Because Julia specifically talks about content design in relation to startups, she focuses on how to achieve the best results possible with a small, lean team. To that end, Julia discusses four must-haves for content design:

  1. Voice
  2. Tools for efficiency
  3. Words in the experience
  4. Ways to check, test, and perfect your words

Voice 🎙️

Why is your company’s voice important? Voice tells your users who you are, creates meaningful connections, and provides valuable signals that convey whether or not your company is deserving of trust. Choosing the voice for your startup begins with a competitor audit.  Documenting who you compete against, and how you might want to differentiate your startup is crucial to finding your corner of the market. For example, is your voice welcoming, gentle, and positive, or are you more formal and technical? 

User research can also be really helpful when determining and monitoring your voice. Involve your research team and learn what does and doesn’t delight your audience when it comes to your messaging.

It’s also important to map your voice to your startup’s values. Be sure to connect to your mission and your brand. Julia sums up product voice as:

Product voice = your values + space to differentiate + what research tells you

So, when you find your voice, where can you lean into it? There are several key areas or moments that provide opportunities to share your unique voice, such as:

  • Notifications: Emails, SMS, and in-app messages are a great place to delight customers
  • Success states: Celebrate with your users in your voice (and remove any anxiety that may be there)
  • Empty states: They aren’t just a chance to educate, they’re a chance to add some interest or fun (or to mask a UX issue).
  • Placeholder text: If a field is well labeled, you can use this section to bring joy and reduce a user’s anxiety.
  • Onboarding: You never get a second chance to make a first impression. Make it count!

Tools for efficiency ⚒️

To remain lean and efficient as a startup, one of the best things you can do is create a style guide. This helps to keep your content and voice consistent. For example, what pronouns do you use in your interface, do you capitalize certain words, etc? There is actually a lot to consider here, so Julia points viewers to various resources that allow you to copy and paste, such as Quinn Keast’s Product Language Framework.

A glossary or language bank is also important. Record branded words, terms that you never use, and terms that you’ve heard your users say organically. This helps to ensure that you’re using language that resonates with your audience and language that reduces cognitive load as much as possible.

Pro tip: Use the Writer app with Figma. This integration helps to ensure that your style guide is actually used! It includes your style guide and glossary so that you’re being consistent as you work. You can also use the Hemingway app or Grammarly to look out for passive voice, hard-to-parse sentences, and overall readability.

Words in the experience – writing for content design 📝

The first thing Julia points out when approaching writing is the need to be user-focused. This might seem obvious to UX practitioners, but word selection can be nuanced, and subtle changes can be powerful. For example, instead of writing “[Your company] introduces a new feature”, think about how can you change the statement to be more about what the new feature means for the user, rather than your company. Here are a few rules of thumb to help refine your writing.

  • Clarity over cleverness. Unless you’re clear and the message is understood by your user, even the best jokes and wittiest phrases in the world will be wasted.
  • Write like you’re having a conversation with your Grandmother. Be clear and don’t use too much jargon.
  • Think like the best content designers. Writing is a process and there are several things to consider, such as the purpose of your copy, the context that it’s being read, and what emotion the reader might be feeling at that moment, etc. Julia offers the Microcopy Canvas as a useful tool for startups, which is a helpful writing template/worksheet created by Jane Ruffino.

Ways to check, test, and perfect your words 👀

Julia suggests that design reviews are the perfect place to sense-check your words and content. Review your designs intentionally and through a content lens. Again, the Microcopy Canvas can be a useful tool when conducting this step, helping to ensure you have considered the right tone and achieved your purpose with your words.

Following a design review process, it’s important to test for clarity and affinity. Conduct user tests frequently to ensure your words and content are clear, understood, and hitting the mark in the intended way.

Finally, make sure your content goals are recorded in your dashboards. Be accountable to your own success measures, KPIs, and OKRs (Objectives and Key Results). Some metrics that help track success are:

  • Onboarding flows
  • Notification metrics
  • Feature adoption
  • Conversion rates

If you’re falling short on some metrics, review your content and try to figure out where words can be sharpened to be clearer, more friendly, or less technical, for example. Then, feed this information into your prioritization and planning. What changes are going to have the most impact on your product’s success? What changes are quick wins? 

Why it matters 🤯

Julia’s talk is important for UX and content designers, particularly those working in startup environments, as it highlights the critical role of content design in achieving product success. The content you share, the voice and tone you adopt, and the clarity of communication, all add to the user's overall experience with your product. Investing time into your content is critical and, as Julia explains, it doesn’t have to put too much stress on your team's workload. If time isn’t invested, however, you may find yourself with poor content, delivering poor experiences, resulting in high customer attrition. 

Efficiency, therefore, should be a focus for startups wanting to achieve great content design without being weighed down. Julia offers pragmatic advice on maintaining consistency through tools like style guides and language banks and by leveraging apps like Hemingway and Grammarly. Tools like these are incredibly helpful when streamlining processes and ensuring a cohesive and polished user interface. 

At the end of the day, Julia stresses the impact that content design has on user experiences and encourages startups to pay close attention to content in ways that are achievable for small teams.

Seeing is believing

Explore our tools and see how Optimal makes gathering insights simple, powerful, and impactful.