January 9, 2020

Anatomy of a Website Footer: Key Elements, UX Best Practices, and Examples

Definition of a website footer

The footer of a website sits at the very bottom of every single web page and contains links to various types of content on your website. It’s an often overlooked component of a website, but it plays several important roles in your information architecture (IA) – it’s not just some extra thing that gets plonked at the bottom of every page.

Getting your website footer right matters!

The footer communicates to your website visitors that they’ve reached the bottom of the page and it’s also a great place to position important content links that don’t belong anywhere else – within reason. A website footer is not a dumping ground for random content links that you couldn’t find a home for, however there are some content types that are conventionally accessed via the footer e.g., privacy policies and copyright information just to name a few.

Lastly, from a usability and navigation perspective, website footers can serve as a bit of a safety net for lost website visitors. Users might be scrolling and scrolling trying to find something and the footer might be what catches them and guides them back to safety before they give up on your website and go elsewhere. Footers are a functional and important part of your overall IA, but also have their own architecture too.

Read on to learn about the types of content links that might be found in a footer, see some real life examples and discuss some approaches that you might take when testing your footer to ensure that your website is supporting your visitors from top to bottom.

What belongs in a website footer

Deciding which content links belong in your footer depends entirely on your website. The type of footer, its intent and content depends on its audience of your customers, potential customers and more — ie your website visitors. Every website is different, but here’s a list of links to content types that might typically be found in a footer.

  • Legal content that may include: Copyright information, disclaimer, privacy policy, terms or use or terms of service – always seek appropriate advice on legal content and where to place it!
  • Your site map
  • Contact details including social media links and live chat or chat bot access
  • Customer service content that may include: shipping and delivery details, order tracking, returns, size guides, pricing if you’re a service and product recall information.
  • Website accessibility details and ways to provide feedback 
  • ‘About Us’ type content that may include: company history, team or leadership team details, the careers page and more 
  • Key navigational links that also appear in the main navigation menu that is presented to website visitors when they first land on the page (e.g. at the top or the side)

Website footer examples

Let’s take a look at three diverse real life examples of website footers.


IKEA US

IKEA’s US website has an interesting double barrelled footer that is also large and complex – a ‘fat footer’ as it’s often called – and its structure changes as you travel deeper into the IA. The below image taken from the IKEA US home page shows two clear blocks of text separated by a blue horizontal line. Above the line we have the heading of ‘All Departments’ with four columns showing product categories and below the line there are seven clear groups of content links covering a broad range of topics including customer service information, links that appear in the top navigation menu and careers. At the very bottom of the footer there are social media links and the copyright information for the website.

An image of IKEA US home page footer on their website, from 2019.
IKEA US home page footer (accessed May 2019)

As expected, IKEA’s overall website IA is quite large, and as a website visitor clicks deeper into the IA, the footer starts to change. On the product category landing pages, the footer is mostly the same but with a new addition of some handy breadcrumbs to aid navigation (see below image).

An image of IKEA US product page footer on their website, from 2019.
IKEA US website footer as it appears on the product category landing page for Textiles & Rugs (accessed May 2019).

When a website visitor travels all the way down to the individual product page level, the footer changes again. In the below image found on the product page for a bath mat, while the blue line and everything below it is still there, the ‘All Departments’ section of the footer has been removed and replaced with non-clickable text on the left hand side that reads as ‘More Bath mats’ and a link on the right hand side that says ‘Go to Bath mats’. Clicking on that link takes the website visitor back to the page above.

IKEA US website footer as it appears on the product page for a bath mat, from 2019.
IKEA US website footer as it appears on the product page for a bath mat (accessed May 2019).

Overall, evolving the footer content as the website visitor progresses deeper into the IA is an interesting approach - as the main page content becomes more focussed as does the footer while still maintaining multiple supportive safety net features.

M.A.C Cosmetics US

The footer for the US website of this well known cosmetics brand has a four part footer. At first it appears to just have three parts as shown in the image below: a wide section with seven content link categories covering a broad range of content types as the main part with a narrow black strip on either end of it making up the second and third parts. The strip above has loyalty program and live chat links and the strip below contains mostly links to legal content.

MAC Cosmetics US website footer with three parts as it appears on the home page upon first glance, from 2019.
MAC Cosmetics US website footer with three parts as it appears on the home page upon first glance (accessed May 2019).


When a website visitor hovers over the ‘Join our loyalty program’ call to action (CTA) in that top narrow strip, the hidden fourth part of the footer which is slightly translucent pulls up like a drawer and sits directly above the strip capping off the top of the main section (as shown in the below image). This section contains more information about the loyalty program and contains further CTAs to join or sign in. It disappears when the cursor is moved away from the hover CTA or it can be collapsed manually via the arrow in the top right hand corner of this fourth part. It’s an interesting and unexpected interaction to have with a footer, but it adds to the overall consistent and cohesive experience of this website because it feels like the footer is an active participant in that experience.

MAC Cosmetics US website footer as it appears on the home page with all four parts visible, from 2019.

MAC Cosmetics US website footer as it appears on the home page with all four parts visible (accessed May 2019).


Domino’s Pizza US

Domino’s Pizza’s US website has a reasonably flat footer in terms of architecture but it occupies as much space as a more complex or deeper footer. As shown in the image below, its content links are presented horizontally over three rows on the left hand side of the footer and these links are visually separated by forward slashes. It also displays social media links and some advertising content on the right hand side. The most interesting feature of this footer is the large paragraph of text titled ‘Legal Stuff’ below the links. Delightfully it uses direct, clear and plain language and even includes a note about delivery charges not including tips and to ‘Please reward your driver for awesomeness’.

Domino’s Pizza US website footer as it appears on the home page, from 2019.

Domino’s Pizza US website footer as it appears on the home page (accessed May 2019).

How to test a website footer

Like every other part of your website, the only way you’re going to know if your footer is supporting your website visitors is if you test it with them. When testing a website’s IA overall, the footer is often excluded. This might be because we want to focus on other areas first or maybe it’s because testing everything at once has the potential to be overwhelming for our research participants.

Testing a footer is fairly easy thing to do and there’s no right or wrong approach – it really does depend on where you are up to in your project, the resources you have available to you and the size and complexity of the footer itself!

If you’re designing a footer for a new website there’s a few ways you might approach ensuring your footer is best supporting your website visitors. If you’re planning to include a large and complex footer, it’s a good idea to start by running an open card sort just on those footer links. An open card sort will help you understand how your website visitors expect those content links in your footer to be grouped and what they think those groups should be called.

If you’re redesigning an existing website, you might first run a tree test on the existing footer to benchmark test it and to pinpoint the exact issues. You might tree test just the footer in the study or you might test the whole website including the footer. Optimal's tree testing is really flexible and you can tree test just a small section of an IA or you can do the whole thing in one go to find out where people are getting lost in the structure. Your approach will depend on your project and what you already know so far. If you suspect there may be issues with the website’s footer, for example, if no one is visiting it and/or you’ve been receiving customer service requests from visitors to help them find content that only lives in the footer,  it would be a good idea to consider isolating it for testing. This will help you avoid any competition between the footer and the rest of your IA as well as any potential confusion that may arise from duplicated tree branches (i.e. when your footer contains duplicate labels).

If you’re short on time and there aren’t any known issues with the footer prior to a redesign, you might tree test the entire IA in your benchmark study, iterate your design and then along with everything else, include testing activities for your footer in your moderated usability testing plan. You might include a usability testing scenario or question that requires your participants to complete a task that involves finding content that can only be found in the footer (e.g., shipping information if it’s an ecommerce website). Also keep a close eye on how your participants are moving around the page in general and see if/when the footer comes into play – is it helping people when they’re lost and scrolling? Or is it going unnoticed? If so, why and so on. Talk to your research participants like you would about any other aspect of your website to find out what’s going on there. When resources are tight, use your best judgement and choose the research approach that’s best for your situation, we’ve all had moments where we’ve had to be pragmatic and do our best with what we have.

When you’re at a stage in your design process where you have a visual design or concept for your footer, you could also run a first-click test. First-click tests are quick and easy and will help you determine how your website visitors are faring once they reach your footer and if they can identify the correct content link to complete their task. Studies can be run remotely or in person and just like the rest of the tools in Optimal's user research platform, are super quick to run and great for reaching website visitors all over the world simply by sharing a link to the study.

Share this article
Author
Optimal
Workshop

Related articles

View all blog articles
Learn more
1 min read

Top Tasks in UX: How to Identify What Really Matters to Your Users

All the way back in 2014, the web passed a pretty significant milestone: 1 billion websites. Of course, fewer than 200 million of these are actually active as of 2019, but there’s an important underlying point. People love to create. If the current digital age that we live in has taught us anything, it’s that it’s never been as easy to get information and ideas out into the world.

Understandably, this ability has been used – and often misused. Overloaded, convoluted websites are par for the course, with a common tactic for website renewal being to simply update them with a new coat of paint while ignoring the swirling pile of outdated and poorly organized content below.

So what are you supposed to do when trying to address this problem on your own website or digital project? Well, there’s a fairly robust technique called top tasks management. Here, we’ll go over exactly what it is and how you can use it.

Getting to grips with top tasks

Ideally, all websites would be given regular, comprehensive reviews. Old content could be revisited and analyzed to see whether it’s still actually serving a purpose. If not, it could be reworked or just removed entirely. Based on research, content creators could add new content to address user needs. Of course, this is just the ideal. The reality is that there’s never really enough time or resource to manage the growing mass of digital content in this way. The solution is to hone in on what your users actually use your website for and tailor the experience accordingly by looking at top tasks.

What are top tasks? They're basically a small set of tasks (typically around 5, but up to 10 is OK too) that are most important to your users. The thinking goes that if you get these core tasks right, your website will be serving the majority of your users and you’ll be more likely to retain them. Ignore top tasks (and any sort of task analysis), and you’ll likely find users leaving your website to find something else that better fits their needs.

The counter to top tasks is tiny tasks. These are everything on a website that’s not all that important for the people actually using it. Commonly, tiny tasks are driven more by the organization’s needs than those of the users. Typically, the more important a task is to a user, the less information there is to support it. On the other hand, the less important a task is to a user, the more information there is. Tiny tasks stem very much from ‘organization first’ thinking, wherein user needs are placed lower on the list of considerations.

According to Jerry McGovern (who penned an excellent write-up of top tasks on A List Apart), the top tasks model says “Focus on what really matters (the top tasks) and defocus on what matters less (the tiny tasks).”

How to identify top tasks

Figuring out your top tasks is an important step in clearing away the fog and identifying what actually matters to your users. We’ll call this stage of the process task discovery, and these are the steps:

  1. Gather: Work with your organization to gather a list of all customer tasks
  2. Refine: Take this list of tasks to a smaller group of stakeholders and work it down into a shortlist
  3. User feedback: Go out to your users and get a representative sample to vote on them
  4. Finalise: Assemble a table of tasks with the one with the highest number of votes at the top and the lowest number of votes at the bottom

We’ll go into detail on the above steps, explaining the best way of handling each one. Keep in mind that this process isn’t something you’ll be able to complete in a week – it’s more likely a 6 to 8-week project, depending on the size of your website, how large your user base is and the receptiveness of your organization to help out.

Step 1: Gather – Figure out the long list of tasks

The first part of the task process is to get out into the wider organization and discover what your users are actually trying to accomplish on your website or by using your products. It’s all about getting into the minds of your users – trying to see the world through their eyes, effectively.

If you’re struggling to think of places where you might find customer tasks, here are some of the best sources:

  • Analytics: Take a deep dive into the analytics of your website or product to find out how people are using them. For websites, you’ll want to look at pages with high traffic and common downloads or interactions. The same applies to products – although the data you have access to will depend on the analytics systems in place.
  • Customer support teams: Your own internal support teams can be a great source of user tasks. Support teams commonly spend all day speaking to users, and as a result, are able to build up a cohesive understanding of the types of tasks users commonly attempt.
  • Sales teams: Similarly, sales teams are another good source of task data. Sales teams typically deal with people before they become your users, but a part of their job is to understand the problems they’re trying to solve and how your website or product can help.
  • Direct customer feedback: Check for surveys your organization has run in the past to see whether any task data already exists.
  • Social media: Head to Twitter, Facebook and LinkedIn to see what people are talking about with regards to your industry. What tasks are being mentioned?

It’s important to note that you need to cast a wide net when gathering task data. You can’t just rely on analytics data. Why? Well, downloads and page visits only reflect what you have, but not what your users might actually be searching for.

As for search, Jerry McGovern explains why it doesn’t actually tell the entire story: “When we worked on the BBC intranet, we found they had a feature called “Top Searches” on their homepage. The problem was that once they published the top searches list, these terms no longer needed to be searched for, so in time a new list of top searches emerged! Similarly, top tasks tend to get bookmarked, so they don’t show up as much in search. And the better the navigation, the more likely the site search is to reflect tiny tasks.”

At the end of the initial task-gathering stage you should be left with around 300 to 500 tasks. Of course, this can scale up or down depending on the size of the website or product.

Step 2: Refine – Create your shortlist

Now that you’ve got your long list of tasks, it’s time to trim them back until you’ve got a shortlist of 100 or less. Keep in mind that working through your long list of tasks is going to take some time, so plan for this process to take at least 4 weeks (but likely more).

It’s important to involve stakeholders from across the organization during the shortlist process. Bring in people from support, sales, product, marketing and leadership areas of the organization. In addition to helping you to create a more concise and usable list, the shortlist process helps your stakeholders to think about areas of overlap and where they may need to work together.

When working your list down to something more usable, try and consolidate and simplify. Stay away from product names as well as internal organization and industry jargon. With your tasks, you essentially want to focus on the underlying thing that a user is trying to do. If you were focusing on tasks for a bank, opt for “Transactions” instead of “Digital mobile payments”. Similarly, bring together tasks where possible. “Customer support”, “Help and support” and “Support center” can all be merged.

At a very technical level, it also helps to avoid lengthy tasks. Stick to around 7 to 8 words and try and avoid verbs, using them only when there’s really no other option. You’ll find that your task list becomes quite to navigate when tasks begin with “look”, “find” and “get”. Finally, stay away from specific audiences and demographics. You want to keep your tasks universal.

Step 3: User feedback – Get users to vote

With your shortlist created, it’s time to take it to your users. Using a survey tool like Optimal's Surveys, add in each one of your shortlisted tasks and have users rank 5 tasks on a scale from 1 to 5, with 5 being the most important and 1 being the least important.

If you’re thinking that your users will never take the time to work through such a long list, consider that the very length of the list means they’ll seek out the tasks that matter to them and ignore the ones that don’t.

A section of the customer survey in Questions.
A section of the customer survey in Questions.

Step 4: Finalize – Analyze your results

Now for the task analysis side of the project. What you want at the end of the user survey end of the project is a league table of entire shortlist of tasks. We’re going to use the example from Cisco’s top tasks project, which has been documented over at A List Apart by Gerry McGovern (who actually ran the project). The entire article is worth a read as it covers the process of running a top task project for a large organization.

Here’s what a league table of the top 20 tasks looks like from Cisco:

A league table of the top 20 tasks from Cisco’s top tasks project.
A league table of the top 20 tasks from Cisco’s top tasks project. Credit: Jerry McGovern.

Here’s the breakdown of the vote for Cisco’s tasks:

  • 3 tasks got the first 25 percent of the vote
  • 6 tasks got 25-50 percent of the vote
  • 14 tasks got 50-75 percent of the vote
  • 44 tasks got 75-100 percent of the vote

While the pattern may seem surprising, it’s actually not unusual. As Jerry explains: “We have done this process over 400 times and the same patterns emerge every single time.”

Final thoughts

Focusing on top tasks management is really a practice that needs to be conducted on a semi-regular basis. The approach benefits organizations in a multitude of ways, bringing different teams and people together to figure out how to best address why your users are coming to your website and what they actually need from you.

As we explained at the beginning of this article, top tasks is really about clearing away the fog and understanding on what really matters. Instead of spreading yourself thin and focusing on a host of tiny tasks, hone in on those top tasks that actually matter to your users.

Understanding how to improve your website

The top tasks approach is an effective way of giving you a clear idea of what you should be focusing on when designing or redesigning your website, but this should really just be one aspect of the work you do.

Utilizing a host of other UX research methods can give you a much more comprehensive idea of what’s working and what’s not. With card sorting, for example, you can learn how your users think the content on your website should be arranged. Then, with this data in hand, you can use tree testing to assemble draft structures of your website and test how people navigate their way through it. You can keep iterating on these structures to ensure you’ve created the most user-friendly navigation.

Take a look at our 101 guides to learn more about card sorting and tree testing, as well as the other user research methods you can use to make solid improvements to your website. If you’d rather just start putting methods into practice using user research tools, take our UX platform for a spin for free here.

Learn more
1 min read

The AI Automation Breakthrough: Key Insights from Our Latest Community Event

Last night, Optimal brought together an incredible community of product leaders and innovators for "The Automation Breakthrough: Workflows for the AI Era" at Q-Branch in Austin, Texas. This two-hour in-person event featured expert perspectives on how AI and automation are transforming the way we work, create, and lead.

The event featured a lightning Talk on "Designing for Interfaces" featured Cindy Brummer, Founder of Standard Beagle Studio, followed by a dynamic panel discussion titled "The Automation Breakthrough" with industry leaders including Joe Meersman (Managing Partner, Gyroscope AI), Carmen Broomes (Head of UX, Handshake), Kasey Randall (Product Design Lead, Posh AI), and Prateek Khare (Head of Product, Amazon). We also had a fireside chat with our CEO, Alex Burke and Stu Smith, Head of Design at Atlassian. 

Here are the key themes and insights that emerged from these conversations:

Trust & Transparency: The Foundation of AI Adoption

Cindy emphasized that trust and transparency aren't just nice-to-haves in the AI era, they're essential. As AI tools become more integrated into our workflows, building systems that users can understand and rely on becomes paramount. This theme set the tone for the entire event, reminding us that technological advancement must go hand-in-hand with ethical considerations.

Automation Liberates Us from Grunt Work

One of the most resonant themes was how AI fundamentally changes what we spend our time on. As Carmen noted, AI reduces the grunt work and tasks we don't want to do, freeing us to focus on what matters most. This isn't about replacing human workers, it's about eliminating the tedious, repetitive tasks that drain our energy and creativity.

Enabling Creativity and Higher-Quality Decision-Making

When automation handles the mundane, something remarkable happens: we gain space for deeper thinking and creativity. The panelists shared powerful examples of this transformation:

Carmen described how AI and workflows help teams get to insights and execution on a much faster scale, rather than drowning in comments and documentation. Prateek encouraged the audience to use automation to get creative about their work, while Kasey shared how AI and automation have helped him develop different approaches to coaching, mentorship, and problem-solving, ultimately helping him grow as a leader.

The decision-making benefits were particularly striking. Prateek explained how AI and automation have helped him be more thoughtful about decisions and make higher-quality choices, while Kasey echoed that these tools have helped him be more creative and deliberate in his approach.

Democratizing Product Development

Perhaps the most exciting shift discussed was how AI is leveling the playing field across organizations. Carmen emphasized the importance of anyone, regardless of their role, being able to get close to their customers. This democratization means that everyone can get involved in UX, think through user needs, and consider the best experience.

The panel explored how roles are blurring in productive ways. Kasey noted that "we're all becoming product builders" and that product managers are becoming more central to conversations. Prateek predicted that teams are going to get smaller and achieve more with less as these tools become more accessible.

Automation also plays a crucial role in iteration, helping teams incorporate customer feedback more effectively, according to Prateek.

Practical Advice for Navigating the AI Era

The panelists didn't just share lofty visions, they offered concrete guidance for professionals navigating this transformation:

Stay perpetually curious. Prateek warned that no acquired knowledge will stay with you for long, so you need to be ready to learn anything at any time.

Embrace experimentation. "Allow your process to misbehave," Prateek advised, encouraging attendees to break from rigid workflows and explore new approaches.

Overcome fear. Carmen urged the audience not to be afraid of bringing in new tools or worrying that AI will take their jobs. The technology is here to augment, not replace.

Just start. Kasey's advice was refreshingly simple: "Just start and do it again." Whether you're experimenting with AI tools or trying "vibe coding," the key is to begin and iterate.

The energy in the room at Q-Branch reflected a community that's not just adapting to change but actively shaping it. The automation breakthrough isn't just about new tools, it's about reimagining how we work, who gets to participate in product development, and what becomes possible when we free ourselves from repetitive tasks.

As we continue to navigate the AI era, events like this remind us that the most valuable insights come from bringing diverse perspectives together. The conversation doesn't end here, it's just beginning.

Interested in joining future Optimal community events? Stay tuned for upcoming gatherings where we'll continue exploring the intersection of design, product, and emerging technologies.

Learn more
1 min read

5 Signs It's Time to Switch Your Research Platform

How to Know When Your Current Tool Is Holding You Back

Your research platform should accelerate insights, not create obstacles. Yet many enterprise research teams are discovering their tools weren't built for the scale, velocity, and quality standards that today’s product development demands.

If you're experiencing any of these five warning signs, it might be time to evaluate alternatives.

1. Your Research Team Is Creating Internal Queues

The Challenge: When platforms limit concurrent studies, research becomes a first-come-first-served bottleneck and urgent research gets delayed by scheduled projects. In fast-moving businesses, research velocity directly impacts competitiveness. Every queued study is a delayed product launch, a missed market opportunity, or a competitor gaining ground.

The Solution: Enterprise-grade research platforms allow unlimited concurrent studies. Multiple teams can research simultaneously without coordination overhead or artificial constraints. Organizations that remove study volume constraints report 3-4x increases in research velocity within the first quarter of switching platforms.

2. Pricing Has Become Unpredictable 

The Problem: When pricing gest too complicated, it becomes unpredictable. Some businesses have per-participant fees, usage caps and seat limits not to mention other hidden charges. Many pricing models weren't designed for enterprise-scale research, they were designed to maximize per-transaction revenue. When you can't predict research costs, you can't plan research roadmaps. Teams start rationing participants, avoiding "expensive" audiences, or excluding stakeholders from platform access to control costs.

The Solution: Transparent, scalable pricing with unlimited seats that grows with your needs.  Volume-based plans that reward research investment rather than penalizing growth. No hidden per-participant markups. 

3. Participant Quality Is Declining

The Problem: This is the most dangerous sign because it corrupts insights at the source. Low-quality participants create low-quality data, which creates poor product decisions.

Warning signs include:

  • Participants using AI assistance during moderated sessions
  • Bot-like response patterns in surveys
  • Participants who clearly don't meet screening criteria
  • Low-effort responses that provide no actionable insight
  • Increasing "throw away this response" rates in your analysis

Poor participant quality isn't just frustrating, it's expensive. Research with the wrong participants produces misleading insights that derail product strategy, waste development resources, and damage market positioning.

The Solution: Multi-layer fraud prevention systems. Behavioral verification. AI-response detection. Real-time quality monitoring. 100% quality guarantees backed by participant replacement policies. When product, design and research teams work with brands that offer 100% participant quality guarantees, they know that they can trust their research and make real business decisions from their insights. 

4. You Can't Reach Your Actual Target Audience

The Problem: Limited panel reach forces compromises. Example: You need B2B software buyers but you get anyone who's used software. Research with "close enough" participants produces insights that don't apply to your actual market. Product decisions based on proxy audiences fail in real-world application.

The solution: Tools like Optimal that offer 10M+ participants across 150+ countries with genuine niche targeting capabilities. Proven Australian market coverage from broad demographics to specialized B2B audiences. Advanced screening beyond basic demographics.

5. Your Platform Hasn't Evolved with Your Needs

The Problem: You chose your platform 3-5 years ago when you were a smaller team with simpler needs. But your organization has grown, research has become more strategic, and your platform's limitations are now organizational constraints. Platform limitations become organizational limitations. When your tools can't support enterprise workflows, your research function can't deliver enterprise value.

The Solution: Complete research lifecycle support from recruitment to analysis. AI-powered insight generation. Enterprise-grade security and compliance. Dedicated support and onboarding. Integration ecosystems that connect research across your organization.

Why Enterprises Are Switching to Optimal

Leading product, design and research teams are moving to Optimal because it's specifically built to address the pain points outlined above:

  1. No Study Volume Constraints: Run unlimited concurrent studies across your entire organization
  2. Transparent, Scalable Pricing: Flexible plans with unlimited seats and predictable costs
  3. Verified Quality Guarantee: 10M+ participants with multi-layer fraud prevention and 100% quality guarantee
  4. Enterprise-Grade Platform: Complete research lifecycle tools, AI-powered insights, dedicated support

Next Steps 

If you're experiencing any of these five signs, it's worth exploring alternatives. The cost of continuing with inadequate tools, delayed launches, poor data quality, limited research capacity, far outweigh the effort of evaluation.

Start a Free Trial – Test Optimal with your real research projects

Compare Platforms – See detailed capability comparisons

Talk to Our Team – Discuss your specific research needs with Australian experts

Seeing is believing

Explore our tools and see how Optimal makes gathering insights simple, powerful, and impactful.