Myna AI

| 2025

Pivoting an AI product from novelty to trusted execution for restaurant owners

Role

Founding Product Designer

Team

CEO, CTO, AI Engineer

Domain

B2B SaaS · AI for Restaurant Tech

Timeline

12 Weeks

TL;DR

Myna is an AI-powered marketing assistant for restaurant owners that helps them respond to reviews, manage social media, and run campaigns through high-value growth actions.


Myna originally shipped as a swipe-based “approve/dismiss” AI assistant. Early testing showed that speed didn’t translate to action or trust.


I led research and rebuilt the product model around outcome-driven tasks, weekly cadence, and constrained AI interactions to increase confidence and follow-through.

Results

Boosted task completion rate

+65%

Reduced user confusion

-70%

Cut down AI-related complaints

-50%

STAGE 1 - UNDERSTANDING WHY THE PRODUCT FAILED

On paper, the experience felt fast and lightweight.

In reality, it wasn’t working.

Early Product Vision

Myna launched as a swipe-based marketing assistant. Owners could swipe right to approve suggested actions or swipe left to dismiss them “one gesture, zero friction.”

Owners didn’t trust the system enough to act, so AI value never landed.

The swipe interaction reduced effort, but also removed meaning, trust, and decision confidence.

Why the Original Experience Failed

When asked to explain the experience, users said:

“What am I looking at?”

“It would make things faster.. but I wouldn’t pay for it.”

“I can just use ChatGPT and get all this done for free."

UX Audit and user test issues mapped on swipe cards

Reviewing usability tests and interaction recordings revealed the same breakdowns across sessions.

The swipe interaction reduced effort, but also removed meaning, trust, and decision confidence.


Reviewing usability tests and interaction recordings revealed the same breakdowns across sessions.

1

The page felt informational instead of task-driven

2

Weak visual hierarchy and no clear primary action

3

Unclear swipe vs scroll cues

4

No context for why actions mattered

What Broke

The issue wasn’t usability alone; it was clarity and trust.

Returning users

<10%

Task Completion rate

0%

Willingness to pay

Low

Evidence

Swipe concept was defined without any user validation.

Misalignment and unclear value

1

Early decisions were assumption-driven, shaped by speed-to-market pressure and investor expectations.

No dedicated UX research

2

As a result, the product optimized for interaction speed instead of confidence in decision-making.

Root Cause

To understand what restaurant owners actually needed, I led a focused research sprint.

Reframe Myna from “idea generator” → “system that guides execution”

Research Summary:


  • Owners are time-poor and decision-fatigued.

  • They prefer guided execution over flexibility.

  • Dashboards, feeds, and swipe interactions failed across all users.

1

User interviews (13 owners)

3

Affinity mapping + behavioral synthesis.

2

Competitive benchmarking + SWOT

4

Quantitative & qualitative analysis

Research Methods:

Research: Reframing the Problem

Rather than designing for an “average user,” synthesis revealed 4 distinct behavioral groups:

The Overwhelmed Operator

Daily operations · Low time · Low tolerance


  • Runs the business day to day; marketing is deprioritized

  • Needs clear, high-impact actions

  • Breaks with dashboards, feeds, and exploration


  • Constantly evaluates if a tool is worth the effort

  • Needs predictability and context before trusting AI

  • Breaks with novelty UX and vague recommendations

The Skeptical Pragmatist

Cautious · ROI-driven · Trust-sensitive


  • Cares about reviews, leads, and revenue

  • Will pay when effort-to-impact is obvious

  • Breaks with shallow engagement

The Outcome-Driven Owner

Results first · Opportunist · Revenue focused


  • Constant context switching

  • Wants the product to think ahead and guide action

  • Breaks with options and interpretation-heavy tools

The Time-Constrained Solo Manager

Single decision-maker · High cognitive load

Key insight:

All groups struggled with feeds, dashboards, and swipe interactions but responded well to clear, outcome-driven tasks.

Research to Product Structure

STAGE 2: REFRAMING THE PROBLEM

"How can I make Myna less of a tool… and more of a partner?"

The real problem was not generating ideas.

It was getting the right work done consistently without requiring ongoing thinking.

This meant shifting Myna from a system that suggested work to one that guided execution.

What problem we were solving

Restaurant owners are:


  • Time-constrained

  • Decision-fatigued

  • Skeptical of AI

"If marketing were reframed as a small set of outcome-driven tasks with proactive guidance, owners would feel more in control, trust the system, and take consistent action."

Hypothesis

System:


  • Decided what mattered

  • Limited how much work appeared

  • And clearly defined what “done” meant

User:


  • Stop evaluating

  • Act with confidence

  • And trust the system more over time.

Easy to Understand

Clear prioritization and hierarchy so users immediately know what matters and why.

Easy to Act On

Decisions should feel lightweight and achievable, even in short sessions.

Trust First, Then Intelligence

AI must feel reliable and predictable before it feels powerful.

Before designing screens, I defined principles to guide how AI, tasks, and decisions should appear for restaurant owners.

Easy to Understand

Clear prioritization and hierarchy so users immediately know what matters and why.

Easy to Act On

Decisions should feel lightweight and achievable, even in short sessions.

Guided, Not Guesswork

Proactive system guidance that guide users instead of asking them to interpret.

Trust First, Then Intelligence

AI must feel reliable and predictable before it feels powerful.

Design Principles

I explored 3 structural approaches. Early reviews and usability feedback revealed clear behavioural differences.

Refined swipe model

Fast, daily execution of the top-priority action.

Felt reactive and transactional

But framed work as isolated reactions

No clear signal of how actions compounded into progress

Actions executed correctly

AI-ranked suggestions feed

More context and options surfaced.

Shifted prioritization back to the user

Required scanning and comparison

Increased hesitation and decision fatigue

Looked intelligent

Task-based workflow

Explicitly defined responsibilities and completion.

Clear priorities and scope

Visible start and end states

Stronger sense of control and progress

Reduced thinking, increased follow-through

Concepts Explored

Direction

Why task-based workflow:

This shift from reactive → proactive reduced cognitive load and consistently drove completion.

The difference was ownership of decisions.


  • Swipe and feeds still required users to mentally manage work.

  • Tasks transferred that responsibility to the system.

Final UI

Progressive disclosure: We revealed complexity only as users showed intent

1

Overview: What matters now

2

Prioritization: Why it matters

3

Execution: How to complete it

How this translated into the final product

However, early versions exposed task details immediately. Users scanned, paused, and hesitated. Sessions were short and often interrupted by real-world work.

So I removed details from the entry point and reduced the home screen to only three things:

Everything else was intentionally hidden.

This helped users orient quickly and move forward without thinking too much.

Entry: What Users See First

Once we committed to a task based model, the next challenge was scope and cadence.

Weekly framing reduced anxiety, gave owners flexibility to act when they had time, and created a clearer sense of completion.

Before

Daily tasks made the product feel like something you had to check every day. That framed Myna as a stream of obligations. Missing a day felt like failure, even when the business impact was minimal.

I grouped actions into weekly task cycles. By intentionally limiting how much work appeared at once, users could finish what mattered, feel done, and move on with confidence.

After

Narrowing the System

Old (Swipe-based)

  • Pull-based (user has to look at content)

  • Gamified gestures

  • Shallow actions (“approve content”)

  • No clarity on value

  • No guidance

New Concept (Task-focused)

  • Push-based (app highlights what matters)

  • Structured tasks & flows

  • High-value actions (reviews, metrics, leads)

  • Direct business outcomes

  • Proactive intelligence

What Changed

We piloted the redesigned experience with 10 restaurants in 2-week cohorts.

Patterns emerged quickly:

Owners completed weekly tasks within 1-2 days.

Guided chat felt natural and supportive.

The experience felt focused, not overwhelming.

At the same time, testing surfaced important weaknesses:

Campaigns still felt heavy for some users.

And more critically, early AI behavior wasn’t reliable enough.

Free-form prompts led to vague questions, generic responses, and occasional hallucinations; which quickly eroded trust during early adoption.

See usability comments here

Testing in the Real World

This reduced flexibility; but dramatically increased predictability and confidence.

AI interaction was constrained. Free-form chat was replaced with tap-based interaction.

Before: with chat input

After: tap-based flow replaces chat

While detailed A/B testing or extended usability studies are ongoing, early testing and direct feedback from users helped inform rapid iterations post-pivot.

AI Design Trade-off: Flexibility vs Trust

Boosted task completion rate

+65%

Reduced user confusion

-70%

Cut down AI-related complaints

-50%

Early results showed meaningful improvement in behavior and trust.

We intentionally validated engagement and execution first, before optimizing for revenue.

Metric


Baseline (Old System)


After Redesign (Week 1-4)


Change


How we measured

Task Completion (Adoption)


Less than 20% of recommended actions completed (6 pilot users, Usability test)


85% completed (10 pilot users)


+65% points


Analytics and task logs.

User Confusion (Clarity)


Avg. 3 navigation errors per session


< 2 errors per session.


Confusion: ↓70%


Usability testing and session recordings.

AI-Related Complaints (Trust)


Almost everyone had issues trusting the AI insights.


5/10 users complained about AI hallucinations and generic insights.


Complaints dropped post iteration: ↓50%


Direct user feedback and internal testing.

Continued Engagement (Retention)


N/A (new product pivot)


Drop after Week 1; users completed tasks within 1–2 days


-


Activity logs and follow-up interviews.

Session Duration (Activity)


-


< 3–4 min avg.


↓ (needs engagement loops)


Analytics tracking session duration

Impact

1

Delivered key screens and edge cases rapidly

2

Guided junior designers through structured QA

3

Leveraged Claude code for vibecoding

4

85% of final screens met quality standards

5

Detailed handoff in Figma and QA tracked in GitHub

Execution Under Constraints

On a tight timeline, I focused on delivering high-quality key screens rapidly

Detailed dev handoff in Figma.

QA documented in Figma, issues tracked in GitHub.

As a lean team, each trade-off prioritized speed, clarity, and core usability over visual delight aligning user needs with business goals.

AI Breadth vs. Task Completion

  • Choice: Focused on a small set of high-value workflows instead of a broad AI assistant.

  • Impact: Clearer value and higher completion.

Conversational Flexibility vs. AI Reliability

  • Choice: Constrained chat, temporarily replacing free-text input with tap-based actions to stabilize trust.

  • Impact: Predictable AI behavior and confidence.

Novelty vs. Business Outcomes

  • Choice: Replaced swipe interactions with outcome-focused tasks.

  • Impact: Task completion increased by 65%, user confusion dropped 70%, and willingness to act improved.

Constraints & Design Trade-offs

My Learnings

This project humbled me in the best way possible. It was messy and ambiguous, failed, experimented and learnt fast.

01

Solve real problems first. Design polish cannot replace weak value.

02

Validate early. Observing actual behavior is more reliable than assumptions.

03

Build trust before engagement. Users return for usefulness, not novelty.

04

Prioritize reliability over flashy features. Stable, context-aware AI drives adoption.

06

Failure is a mirror, not a verdict. Every misstep taught more than success, improving the product and my user-centered approach.

06

Failure is a mirror, not a verdict. Every misstep taught more than success, improving the product and my user-centered approach.

Next Steps

Evolve from a task-based workflow to a multi-modal experience combining tasks, insights, content, and guidance

1

Reopen pilots after core improvements to gather deeper qualitative feedback on real business outcomes

2

Validate willingness to pay by identifying the single workflow users value most

3

Strengthen AI trust through improved accuracy and clearer trust signals

4

Unify information architecture across chat, tasks, insights, and notifications

5

Build retention loops tied to real restaurant activity, not vanity engagement

6

Track metrics like adoption, retention, revenue, referrals

See ongoing exploration

I'm currently open for new and exciting opportunities.

Let's connect and create something nice.

I'm glad you made it here.

20:47:51

|

V.2026

Created by Pavan Suresh

Create a free website with Framer, the website builder loved by startups, designers and agencies.