Logan & Friends

| 2024

Delivered near real-time instructional coaching and cut feedback navigation time 40% without letting AI override human judgment

Impact

Efficiency rate

100%

Coaching Feedback Moved From Delayed → Near Real-Time

Eliminated multi-day feedback delays, enabling teachers to reflect while lessons were still fresh.

User satisfaction

90%

90% educator satisfaction with clarity and speed

Trust increased when AI insights were contextual, optional, and reviewed by a human coach.

Navigation time

-40%

Reduced feedback navigation time by 40% (6 min → 3.5 min)

Teachers spent less time searching and more time reflecting and improving potentially.

Context

Logan & Friends is an education company that partners with schools to design equity-focused, real-world learning experiences. A core part of their work involves instructional coaching, observing classroom teaching and providing feedback to help educators grow.

Instructional coaching works in theory.
In practice, it doesn’t scale well.

Feedback often arrives days or weeks after a lesson, long after its impact has faded.

The client’s vision was ambitious:

to explore whether AI could reduce feedback delays by analyzing voice-recorded teaching sessions, generating standards-aligned insights, and supporting human coaches in delivering faster, more actionable guidance.

However, the scope, technical feasibility, and trust boundaries of using AI in classrooms were unclear.

TL;DR

Instructional coaching doesn’t fail because feedback is bad.
It fails because it arrives too late, without context, and without control.

As one of the product designers on the team, I helped design an AI-assisted coaching system that prioritized timeliness, clarity, and human judgment over automation.

By anchoring AI insights to real classroom moments and tightly constraining when and how AI intervened, we reduced navigation friction, eliminated coaching delays, and preserved trust.

Role

Product Designer

Team

4 Product Designers · Client Designer (PhD) · Subject Matter Experts

Domain

EdTech · Human-Centered AI

Timeline

30 Weeks

Problem

The surface problem was delayed feedback.


The deeper problem was misalignment between how coaching tools are designed and how teachers actually reflect and improve.

What wasn’t working


  • Feedback often arrived 72+ hours after observation

  • Guidance felt generic and disconnected from real classroom moments

  • Teachers struggled to review long, dense feedback within busy schedules

  • Existing tools added cognitive load instead of reducing it

At the same time, introducing AI added new risks:


  • Teachers feared being judged or evaluated by algorithms

  • Recording classrooms raised privacy and consent concerns

  • Automation risked removing the empathy central to good coaching

The challenge wasn’t whether AI could generate feedback.

It was whether it could do so without breaking trust.

Research

To ground the product in real educator needs, I led and synthesized research across multiple methods.

Research methods


  • 5 in-depth teacher interviews

  • 12 educator survey responses

  • 7 competitor analyses

  • Review of 20+ research papers on instructional coaching and classroom observation

  • Multiple rounds of usability testing on early concepts

Key research insights

1

Timeliness matters more than volume

Teachers consistently said feedback loses value when it arrives days later, even if it’s detailed.

2

Context builds

trust

Educators were far more receptive to feedback when it was tied to specific classroom moments they could recall.

3

AI acceptance is conditional

Most teachers were open to AI-assisted feedback only if it didn’t feel judgmental, intrusive, or opaque.

4

Navigation friction kills adoption

Usability testing showed teachers spent excessive time finding relevant feedback, increasing frustration and disengagement.

Empathy and affinity mapping reinforced a clear pattern:

Teachers wanted bite-sized, contextual guidance that respected their time and professional judgment.

The Guiding Principle

AI should surface insight, not deliver judgment.

This principle guided every design decision.


  • AI’s role: listen, identify patterns, surface moments worth attention

  • Human coaches’ role: contextualize, interpret, and guide growth

Early Direction: Insight Summaries

I designed the feedback screen to give teachers a clear snapshot of summarized session insights into categorized cards.

Feedback is grouped into clear categories with color-coded tags (1) that highlight framework patterns, with simple, action-focused notes (2), making it feel like guidance rather than evaluation. and a “View related resources” link (3) provides deeper context and help without cluttering the view, keeping feedback approachable and scannable.

What worked


  • Clear structure

  • Fast overview

What failed


  • Teachers couldn’t tell why feedback appeared

  • No connection to specific classroom moments

  • Feedback felt generic, the same frustration teachers already had

Decision 1:

Anchor Feedback to Real Classroom Moments

Based on research and testing, I redesigned feedback to be

time-based and contextual.

  • Switched to a vertical, chronological layout

  • Added timestamps tied directly to the lesson

  • Renamed “Session Insights” to "Action Items"

That way, teachers can instantly see what each note refers to, why it was flagged and implement those actions in practice.

Before

After

Anchoring feedback to real moments made AI feel observational, not evaluative.

This was the first major trust unlock.

Decision 2

Despite improved context, testing showed teachers spent over 6+ minutes navigating feedback.

So I redesigned the feedback experience to be discoverable, scannable first, deep second.

  • Added an audio player with timeline markers

  • Allowed quick jumps to relevant moments

Why it mattered

Teachers are time-poor. If value isn’t easy to find, intelligence doesn’t matter.

Decision 3

Make AI Optional, Contextual, and User-Invoked

Early research showed teachers were open to AI, but only if it didn’t interrupt or judge them.

What changed

  • AI chat was added as an on-demand assistant

  • It responded only to questions about selected feedback

  • It never surfaced unsolicited evaluations or scoring

Why it mattered

AI became a support tool, not an authority. Teachers stayed in control of when, how, and how much they engaged with AI.

This preserved the human core of coaching while still gaining speed.

The Final Experience

The final prototype combined:


  • Timestamped feedback tied to real classroom moments

  • Audio playback with visual markers

  • Clear, scannable feedback cards

  • Optional AI support for clarification and exploration

  • Human coach review before delivery

The system felt less like an evaluation tool and more like a coaching partner.

Validation & Testing

We validated clarity, efficiency, and trust through:


  • Expert reviews with instructional coaches

  • Think-aloud sessions with experienced teachers, novice teachers, and coaches

  • Cognitive walkthroughs and design critiques

  • Post-test satisfaction questionnaires

Key findings and design responses

Issue:

The Audio player remained fixed on top that reduced the usable scroll area, constrained scrolling made longer feedback hard to review.

Before

Fixed:

I reworked the layout to prioritize reading and flow by resizing the existnig containers, expanding the primary scroll area to allow continuous reading and made the audio player sticky.

After

Execution

I led end-to-end design from concept to validation:


  • Defined information architecture and feedback flows

  • Designed low- to mid/high-fidelity prototypes

  • Ran multiple rounds of usability testing

  • Iterated based on observed behavior, not assumptions

  • Collaborated closely with designers, a PhD client, and subject matter experts to ensure pedagogical and technical feasibility

Design decisions were continuously tested and refined to balance clarity, trust, and speed.

Impact

Usability testing and prototype validation showed measurable improvements:

40% reduction in navigation time (6 min → 3.5 min)

Near real-time feedback, eliminating multi-day coaching delays

90% of teachers reported satisfaction with clarity and speed

Increased confidence in AI-assisted feedback when paired with human review

Most importantly, teachers described the experience as supportive, not judgmental.

AI Design Trade-offs

We intentionally traded:


  • Flexibility for trust

  • Automation for clarity

  • Intelligence for restraint

AI worked best when it:


  • stayed close to real moments

  • avoided abstract scoring

  • respected human authority

My Learnings

This project reinforced that designing AI systems is as much about restraint as capability.

1

Trust is earned through predictability, not novelty

2

AI adoption depends on timing, context, and control

3

Human judgment must remain visible in high-stakes domains

4

Iteration is how ambiguity becomes clarity

What’s Next

1

Expand coaching insights across multiple sessions

2

Strengthen privacy and consent controls

3

Validate long-term impact on teacher growth

4

Explore real classroom pilots

I'm currently open for new and exciting opportunities.

Let's connect and create something nice.

I'm glad you made it here.

23:26:02

|

V.2026

Created by Pavan Suresh

Create a free website with Framer, the website builder loved by startups, designers and agencies.