October 7, 2025
3 minutes

The AI Automation Breakthrough: Key Insights from Our Latest Community Event

Last night, Optimal brought together an incredible community of product leaders and innovators for "The Automation Breakthrough: Workflows for the AI Era" at Q-Branch in Austin, Texas. This two-hour in-person event featured expert perspectives on how AI and automation are transforming the way we work, create, and lead.

The event featured a lightning Talk on "Designing for Interfaces" featured Cindy Brummer, Founder of Standard Beagle Studio, followed by a dynamic panel discussion titled "The Automation Breakthrough" with industry leaders including Joe Meersman (Managing Partner, Gyroscope AI), Carmen Broomes (Head of UX, Handshake), Kasey Randall (Product Design Lead, Posh AI), and Prateek Khare (Head of Product, Amazon). We also had a fireside chat with our CEO, Alex Burke and Stu Smith, Head of Design at Atlassian. 

Here are the key themes and insights that emerged from these conversations:

Trust & Transparency: The Foundation of AI Adoption

Cindy emphasized that trust and transparency aren't just nice-to-haves in the AI era, they're essential. As AI tools become more integrated into our workflows, building systems that users can understand and rely on becomes paramount. This theme set the tone for the entire event, reminding us that technological advancement must go hand-in-hand with ethical considerations.

Automation Liberates Us from Grunt Work

One of the most resonant themes was how AI fundamentally changes what we spend our time on. As Carmen noted, AI reduces the grunt work and tasks we don't want to do, freeing us to focus on what matters most. This isn't about replacing human workers, it's about eliminating the tedious, repetitive tasks that drain our energy and creativity.

Enabling Creativity and Higher-Quality Decision-Making

When automation handles the mundane, something remarkable happens: we gain space for deeper thinking and creativity. The panelists shared powerful examples of this transformation:

Carmen described how AI and workflows help teams get to insights and execution on a much faster scale, rather than drowning in comments and documentation. Prateek encouraged the audience to use automation to get creative about their work, while Kasey shared how AI and automation have helped him develop different approaches to coaching, mentorship, and problem-solving, ultimately helping him grow as a leader.

The decision-making benefits were particularly striking. Prateek explained how AI and automation have helped him be more thoughtful about decisions and make higher-quality choices, while Kasey echoed that these tools have helped him be more creative and deliberate in his approach.

Democratizing Product Development

Perhaps the most exciting shift discussed was how AI is leveling the playing field across organizations. Carmen emphasized the importance of anyone, regardless of their role, being able to get close to their customers. This democratization means that everyone can get involved in UX, think through user needs, and consider the best experience.

The panel explored how roles are blurring in productive ways. Kasey noted that "we're all becoming product builders" and that product managers are becoming more central to conversations. Prateek predicted that teams are going to get smaller and achieve more with less as these tools become more accessible.

Automation also plays a crucial role in iteration, helping teams incorporate customer feedback more effectively, according to Prateek.

Practical Advice for Navigating the AI Era

The panelists didn't just share lofty visions, they offered concrete guidance for professionals navigating this transformation:

Stay perpetually curious. Prateek warned that no acquired knowledge will stay with you for long, so you need to be ready to learn anything at any time.

Embrace experimentation. "Allow your process to misbehave," Prateek advised, encouraging attendees to break from rigid workflows and explore new approaches.

Overcome fear. Carmen urged the audience not to be afraid of bringing in new tools or worrying that AI will take their jobs. The technology is here to augment, not replace.

Just start. Kasey's advice was refreshingly simple: "Just start and do it again." Whether you're experimenting with AI tools or trying "vibe coding," the key is to begin and iterate.

The energy in the room at Q-Branch reflected a community that's not just adapting to change but actively shaping it. The automation breakthrough isn't just about new tools, it's about reimagining how we work, who gets to participate in product development, and what becomes possible when we free ourselves from repetitive tasks.

As we continue to navigate the AI era, events like this remind us that the most valuable insights come from bringing diverse perspectives together. The conversation doesn't end here, it's just beginning.

Interested in joining future Optimal community events? Stay tuned for upcoming gatherings where we'll continue exploring the intersection of design, product, and emerging technologies.

Share this article
Author
Optimal
Workshop
Topics

Related articles

View all blog articles
Learn more
1 min read

When Everyone's a Researcher and it's a Good Thing

Be honest. Are you guilty of being a gatekeeper? 

For years, UX teams have treated research as a specialized skill that requires extensive training, advanced degrees, and membership in the researcher club. We’re guilty of it too! We've insisted that only "real researchers" can talk to users, conduct studies, or generate insights.

But the problem with this is, this gatekeeping is holding back product development, limiting insights, and ironically, making research less effective.  As a result,  product and design teams are starting to do their own research, bypassing UX because they want to just get things done. 

This shift is happening, and while we could view this as the downfall of traditional UX, we see it more as an evolution. And when done right, with support from UX, this democratization actually leads to better products, more research-informed organizations, and yes, more valuable research roles.

The Problem with Gatekeeping 

Product teams need insights constantly, making decisions daily about features, designs, and priorities. Yet dedicated researchers are outnumbered, often supporting 15-20 product team members each. The math just doesn't work. No matter how talented or efficient researchers are, they can't be everywhere at once, answering every question in real-time. This mismatch between insight demand and research capacity forces teams into an impossible choice: wait for formal research and miss critical decision windows or move forward without insights and risk building the wrong thing.

Since product teams often don’t have the time to wait, teams make decisions anyway, without research. A Forrester study found that 73% of product decisions happen without any user input, not because teams don't value research, but because they can't wait weeks for formal research cycles.

In organizations where this is already happening (it’s most of them!) teams have two choices, accept that their research to insight to development workflow is broken, or accept that things need to change and embrace the new era of research democratization. 

In Support of  Research Democratization

The most research-informed organizations aren't those with the most researchers, they're those where research skills are distributed throughout the team. When Product Managers and Designers talk directly to users, with researchers providing frameworks and quality control they make more research-informed decisions which result in better product performance and lower business risk. 

When PMs and designers conduct their own research, context doesn't get lost in translation. They hear the user's words, see their frustrations, and understand nuances that don't survive summarization. But there is a right way to democratize, which not all organizations are doing. 

Democratization as a consequence instead of as an intentional strategy, is chaos. Without frameworks and support from experienced researchers, it just won’t work. The goal isn't to turn everyone into researchers, it's to empower more teams to do their own research, while maintaining quality and rigor. In this model, the researcher becomes an advisor instead of a gatekeeper and the researcher's role evolves from conducting all studies to enabling teams to conduct their own. 

Not all questions need expert researchers. Intercom uses a three-tier model:

  • Tier 1 (70% of questions): Teams handle with proven templates
  • Tier 2 (20% of questions): Researcher-supported team execution
  • Tier 3 (10% of questions): Researcher-led complex studies

This model increased research output by 300% while improving quality scores by 25%.

In a model like this, the researcher becomes more important than ever because democratization needs quality assurance. 

Elevating the Role of Researchers 

Democratization requires researchers to shift from "protectors of methodology" to "enablers of insight." It means:

  • Not seeking perfection because an imperfect study done today beats a perfect study done never.
  • Acknowledging that 80% confidence on 100% of decisions beats 100% confidence on 20% of decisions.
  • Measuring success by the "number of research-informed decisions made” instea dof the "number of studies conducted" 
  • Deciding that more research happening is good, even if researchers aren't doing it all.

By enabling teams to handle routine research, professional researchers focus on:

  • Complex, strategic research that requires deep expertise
  • Building research capabilities across the organization
  • Ensuring research quality and methodology standards
  • Connecting insights across teams and products
  • Driving research-informed culture change

In truly research-informed organizations, everyone has user conversations. PMs do quick validation calls. Designers run lightweight usability tests. Engineers observe user sessions. Customer success shares user feedback.

And researchers? They design the systems, ensure quality, tackle complex questions, and turn this distributed insight into strategic direction.

Research democratization isn't about devaluing research expertise, it's about scaling research impact. It's recognizing that in today's product development pace, the choice isn't between formal research and democratized research. It's between democratized research and no research at all.

Done right, democratization isn't the end of UX research as a profession. It's the beginning of research as a competitive advantage.

Learn more
1 min read

The Great Debate: Speed vs. Rigor in Modern UX Research

Most product teams treat UX research as something that happens to them:  a necessary evil that slows things down or a luxury they can't afford. The best product teams flip this narrative completely. Their research doesn't interrupt their roadmap; it powers it.

"We need insights by Friday."

"Proper research takes at least three weeks."

This conversation happens in product teams everywhere, creating an eternal tension between the need for speed and the demands of rigor. But what if this debate is based on a false choice?

Research that Moves at the Speed of Product

Product development has accelerated dramatically. Two-week sprints are standard. Daily deployment is common. Feature flags allow instant iterations. In this environment, a four-week research study feels like asking a Formula 1 race car to wait for a horse-drawn carriage.

The pressure is real. Product teams make dozens of decisions per sprint, about features, designs, priorities, and trade-offs. Waiting weeks for research on each decision simply isn't viable. So teams face an impossible choice: make decisions without insights or slow down dramatically.

As a result, most teams choose speed. They make educated guesses, rely on assumptions, and hope for the best. Then they wonder why features flop and users churn.

The False Dichotomy

The framing of "speed vs. rigor" assumes these are opposing forces. But the best research teams have learned they're not mutually exclusive, they require different approaches for different situations.

We think about research in three buckets, each serving a different strategic purpose:

Discovery: You're exploring a space, building foundational knowledge, understanding thelandscape before you commit to a direction. This is where you uncover the problems worth solving and identify opportunities that weren't obvious from inside your product bubble.

Fine-Tuning: You have a direction but need to nail the specifics. What exactly should this feature do? How should it work? What's the minimum viable version that still delivers value? This research turns broad opportunities into concrete solutions.

Delivery: You're close to shipping and need to iron out the final details: copy, flows, edge cases. This isn't about validating whether you should build it; it's about making sure you build it right.

Every week, our product, design, research and engineering leads review the roadmap together. We look at what's coming and decide which type of research goes where. The principle is simple: If something's already well-shaped, move fast. If it's risky and hard to reverse, invest in deeper research.

How Fast Can Good Research Be?

The answer is: surprisingly fast, when structured correctly! 

For our teams, how deep we go isn't about how much time we have: it's about how much it would hurt to get it wrong. This is a strategic choice that most teams get backwards.

Go deep when the stakes are high, foundational decisions that affect your entire product architecture, things that would be expensive to reverse, moments where you need multiple stakeholders aligned around a shared understanding of the problem.

Move fast when you can afford to be wrong,  incremental improvements to existing flows, things you can change easily based on user feedback, places where you want to ship-learn-adjust in tight loops.

Think of it as portfolio management for your research investment. Save your "big research bets" for the decisions that could set you back months, not days. Use lightweight validation for everything else.

And while good research can be fast, speed isn't always the answer. There are definitely situations where deep research needs to run and it takes time. Save those moments for high stakes investments like repositioning your entire product, entering new markets, or pivoting your business model. But be cautious of research perfectionism which is a risk with deep research. Perfection is the enemy of progress. Your research team shouldn’t be asking "Is this research perfect?" but instead "Is this insight sufficient for the decision at hand?"

The research goal should always be appropriate confidence, not perfect certainty.

The Real Trade-Off

The choice shouldn’t be  speed vs. rigor, it's between:

  • Research that matters (timely, actionable, sufficient confidence)
  • Research that doesn't (perfect methodology, late arrival, irrelevant to decisions)

The best research teams have learned to be ruthlessly pragmatic. They match research effort to decision impact. They deliver "good enough" insights quickly for small decisions and comprehensive insights thoughtfully for big ones.

Speed and rigor aren't enemies. They're partners in a portfolio approach where each decision gets the right level of research investment. The teams winning aren't choosing between speed and rigor—they're choosing the appropriate blend for each situation.

Learn more
1 min read

When AI Meets UX: How to Navigate the Ethical Tightrope

As AI takes on a bigger role in product decision-making and user experience design, ethical concerns are becoming more pressing for product teams. From privacy risks to unintended biases and manipulation, AI raises important questions: How do we balance automation with human responsibility? When should AI make decisions, and when should humans stay in control?

These aren't just theoretical questions they have real consequences for users, businesses, and society. A chatbot that misunderstands cultural nuances, a recommendation engine that reinforces harmful stereotypes, or an AI assistant that collects too much personal data can all cause genuine harm while appearing to improve user experience.

The Ethical Challenges of AI

Privacy & Data Ethics

AI needs personal data to work effectively, which raises serious concerns about transparency, consent, and data stewardship:

  • Data Collection Boundaries – What information is reasonable to collect? Just because we can gather certain data doesn't mean we should.
  • Informed Consent – Do users really understand how their data powers AI experiences? Traditional privacy policies often don't do the job.
  • Data Longevity – How long should AI systems keep user data, and what rights should users have to control or delete this information?
  • Unexpected Insights – AI can draw sensitive conclusions about users that they never explicitly shared, creating privacy concerns beyond traditional data collection.

A 2023 study by the Baymard Institute found that 78% of users were uncomfortable with how much personal data was used for personalized experiences once they understood the full extent of the data collection. Yet only 12% felt adequately informed about these practices through standard disclosures.

Bias & Fairness

AI can amplify existing inequalities if it's not carefully designed and tested with diverse users:

  • Representation Gaps – AI trained on limited datasets often performs poorly for underrepresented groups.
  • Algorithmic Discrimination – Systems might unintentionally discriminate based on protected characteristics like race, gender, or disability status.
  • Performance Disparities – AI-powered interfaces may work well for some users while creating significant barriers for others.
  • Reinforcement of Stereotypes – Recommendation systems can reinforce harmful stereotypes or create echo chambers.

Recent research from Stanford's Human-Centered AI Institute revealed that AI-driven interfaces created 2.6 times more usability issues for older adults and 3.2 times more issues for users with disabilities compared to general populations, a gap that often goes undetected without specific testing for these groups.

User Autonomy & Agency

Over-reliance on AI-driven suggestions may limit user freedom and sense of control:

  • Choice Architecture – AI systems can nudge users toward certain decisions, raising questions about manipulation versus assistance.
  • Dependency Concerns – As users rely more on AI recommendations, they may lose skills or confidence in making independent judgments.
  • Transparency of Influence – Users often don't recognize when their choices are being shaped by algorithms.
  • Right to Human Interaction – In critical situations, users may prefer or need human support rather than AI assistance.

A longitudinal study by the University of Amsterdam found that users of AI-powered decision-making tools showed decreased confidence in their own judgment over time, especially in areas where they had limited expertise.

Accessibility & Digital Divide

AI-powered interfaces may create new barriers:

  • Technology Requirements – Advanced AI features often require newer devices or faster internet connections.
  • Learning Curves – Novel AI interfaces may be particularly challenging for certain user groups to learn.
  • Voice and Language Barriers – Voice-based AI often struggles with accents, dialects, and non-native speakers.
  • Cognitive Load – AI that behaves unpredictably can increase cognitive burden for users.

Accountability & Transparency

Who's responsible when AI makes mistakes or causes harm?

  • Explainability – Can users understand why an AI system made a particular recommendation or decision?
  • Appeal Mechanisms – Do users have recourse when AI systems make errors?
  • Responsibility Attribution – Is it the designer, developer, or organization that bears responsibility for AI outcomes?
  • Audit Trails – How can we verify that AI systems are functioning as intended?

How Product Owners Can Champion Ethical AI Through UX

At Optimal, we advocate for research-driven AI development that puts human needs and ethical considerations at the center of the design process. Here's how UX research can help:

User-Centered Testing for AI Systems

AI-powered experiences must be tested with real users to identify potential ethical issues:

  • Longitudinal Studies – Track how AI influences user behavior and autonomy over time.
  • Diverse Testing Scenarios – Test AI under various conditions to identify edge cases where ethical issues might emerge.
  • Multi-Method Approaches – Combine quantitative metrics with qualitative insights to understand the full impact of AI features.
  • Ethical Impact Assessment – Develop frameworks specifically designed to evaluate the ethical dimensions of AI experiences.

Inclusive Research Practices

Ensuring diverse user participation helps prevent bias and ensures AI works for everyone:

  • Representation in Research Panels – Include participants from various demographic groups, ability levels, and socioeconomic backgrounds.
  • Contextual Research – Study how AI interfaces perform in real-world environments, not just controlled settings.
  • Cultural Sensitivity – Test AI across different cultural contexts to identify potential misalignments.
  • Intersectional Analysis – Consider how various aspects of identity might interact to create unique challenges for certain users.

Transparency in AI Decision-Making

UX teams should investigate how users perceive AI-driven recommendations:

  • Mental Model Testing – Do users understand how and why AI is making certain recommendations?
  • Disclosure Design – Develop and test effective ways to communicate how AI is using data and making decisions.
  • Trust Research – Investigate what factors influence user trust in AI systems and how this affects experience.
  • Control Mechanisms – Design and test interfaces that give users appropriate control over AI behavior.

The Path Forward: Responsible Innovation

As AI becomes more sophisticated and pervasive in UX design, the ethical stakes will only increase. However, this doesn't mean we should abandon AI-powered innovations. Instead, we need to embrace responsible innovation that considers ethical implications from the start rather than as an afterthought.

AI should enhance human decision-making, not replace it. Through continuous UX research focused not just on usability but on broader human impact, we can ensure AI-driven experiences remain ethical, inclusive, user-friendly, and truly beneficial.

The most successful AI implementations will be those that augment human capabilities while respecting human autonomy, providing assistance without creating dependency, offering personalization without compromising privacy, and enhancing experiences without reinforcing biases.

A Product Owner's Responsibility: Leading the Charge for Ethical AI

As UX professionals, we have both the opportunity and responsibility to shape how AI is integrated into the products people use daily. This requires us to:

  • Advocate for ethical considerations in product requirements and design processes
  • Develop new research methods specifically designed to evaluate AI ethics
  • Collaborate across disciplines with data scientists, ethicists, and domain experts
  • Educate stakeholders about the importance of ethical AI design
  • Amplify diverse perspectives in all stages of AI development

By embracing these responsibilities, we can help ensure that AI serves as a force for positive change in user experience enhancing human capabilities while respecting human values, autonomy, and diversity.

The future of AI in UX isn't just about what's technologically possible; it's about what's ethically responsible. Through thoughtful research, inclusive design practices, and a commitment to human-centered values, we can navigate this complex landscape and create AI experiences that truly benefit everyone.

Seeing is believing

Explore our tools and see how Optimal makes gathering insights simple, powerful, and impactful.