TIMELINE

Sept 2025 - Dec 2025

ROLE

Product Designer

TEAM

CEO (Aude.ai)

Product Manager (Aude.ai)

7 Designers

TLDR

Aude.ai surfaces AI-powered performance insights for engineering teams, but users didn't trust what it was showing them and struggled to act on those insights together. Designed experiences to make AI-generated insights transparent and actionable, and introduced a coaching feature to bring human connection back to performance evaluation.

IMPACT

38% Increase

in perceived transparency of the platform

14% Increase

task success rate

Context

What is Aude.ai?

Aude is an AI-powered performance evaluation platform that integrates with tools like GitHub, Jira, and Slack to capture work patterns and deliver personalized, actionable insights for engineering teams.

Engineers are the primary recipients of this value, using the platform to understand their own performance, get feedback, and improve.

!

Despite this, Aude noticed that adoption among engineers remained low and the platform wasn't being used the way it was intended.

understanding the problem

What was going wrong?

We conducted 8 in-depth interviews with software engineers, 2 with engineering managers, and a digital ethnographic study across Reddit. Three themes kept coming up.

1

AI Insights felt like a black box

Engineers couldn't see where data came from or why suggestions were made, making it hard to trust anything the platform surfaced.

2

Disconnect from the real world

Performance evaluation involves teammates and cross-team collaborators, but the platform had no way to capture or surface that input.

3

AI suggestions lead nowhere

The platform told engineers what to improve but gave them no infrastructure to act on it, leaving them without support to actually grow.

How might we make performance insights more transparent, actionable, and meaningful to an engineer's growth?

challenge 01

Engineers had no reason to trust what they were seeing

Engineers were receiving AI-generated suggestions but had no visibility into what data they were based on or why they were being surfaced. Without that context, suggestions usually felt arbitrary and something you couldn't interrogate, you couldn't trust.

Suggestions

Add clear context and examples in your PR descriptions to speed up reviews. This helps reviewers understand the changes without needing extra clarification

Goal: Engineering Excellence | Criteria: Code Quality & Review Efficiency

Add as Action

Generated: 10/19

In 3 of your recent PRs, edge case failures were caught post-merge by teammates. Proactively flagging known risks in your PR descriptions can prevent downstream issues and demonstrate ownership of quality.

Goal: Trust & Integrity | Criteria: Integrity & Prioritizing the Right Outcome

Add as Action

Generated: 10/19

You've had 4 async threads in the last month where decisions were made without documenting the reasoning. Sharing a brief rationale keeps your team aligned and reduces re-litigating decisions later

Goal: Trust & Integrity | Criteria: Building Team Trust

Add as Action

Generated: 10/19

For me to be okay with it, I'd like to know exactly the reason behind this… you have some black box judging you.

I personally would love some insights, but I need to figure out what I’m supposed to do with them.

The suggestions feel vague. I can’t tell where they’re coming from.

So, We redesigned the suggestions card

to show engineers exactly where data was pulled from, why the suggestion was generated, and what to do next.

challenge 02

Aude's evaluation flow had one blind spot, It only captured what a manager could see.

Engineers working across teams, in code reviews, in Slack threads, in cross-functional projects, weren't getting that visibility surfaced anywhere.

What if the most useful feedback wasn't coming from your manager or an AI, but from your teammates?

— Engineering manager

Research told us what we were already sensing: peer feedback tends to be more accurate than manager-only reviews, but most platforms had no real way to capture or surface it when it matters.

The feedback was always there. We just made it visible.

We designed a view where AI combs through all the peer feedback, finds what keeps coming up, and shows it to the engineer. Bigger circle means more people flagged it. Each one has steps attached so they know where to start.

Patterns recognized from feedback

AI summarized

4 themes derived from 40 mentions across 11 reviewers. Updated 2 hours ago.

Size proportional to mentions

Share updates earlier in the sprint

Improve documentation on PRs

Follow consistent code formatting

Communicate blockers sooner

Selected Theme

Share updates earlier in the sprint

Teammates consistently mention that progress visibility drops mid-sprint. They want earlier signals, not end-of-week writeups.

Actions based on this feedback

Initiate Coaching

Break larger tasks into smaller tickets and share what you will pick up first.

Post a Tuesday mid-sprint update with blockers, wins, and next 48h.

Bring up blockers during daily standup instead of waiting until the end of the week.

challenge 03

!

Coaching was a gap even our project sponsors named upfront. They knew it needed to exist on the platform but didn't know what it should look like.

Our interviews told us two important things

1

Engineers never asked for coaching directly, but mentioned that they would seek help from their manager when they weren't sure how to act on a suggestion.

2

Managers described their 1:1s as where real growth happened, not in any tool.

So instead of trying to build an AI coach, We built a space for humans to coach each other.

Engineers can initiate a coaching session directly from an AI suggestion or a peer feedback pattern they want to work on. That request goes to their manager, who accepts it and schedules the first call.

The manager accepts the coaching request

They get on their first call and align on the challenge

Together they set the milestones, tasks, and a plan to work toward

Coaching

Coaching Session with

Clark

This board is shared between you and Clark

Settings

Current Challenge

Breaking large features into smaller pieces

Description

I understand I need to plan features more carefully before writing code, but I'm unsure what that actually looks like in practice. My last two features needed mid-sprint rewrites because the structure wasn't clear early on, and I want to figure out a consistent approach that works before I start the next one.

Milestones

Align on what "ready to implement" looks like for Clark’s features

Write a pre-implementation breakdown for the next assigned feature

Consistently explain scope and structure in PR descriptions before requesting review

Upcoming Meeting Agenda

  • Walk through Clark’s breakdown of upcoming feature work.

  • Review the pre-implementation outline he drafted this week.

  • Check if the format is sustainable and adjust if needed.

  • Discuss how to flag scope creep early before it causes rewrites.

Summary of the Nov 12, 10 AM meeting

  • Reviewed two recent features where structure broke down mid-implementation.

  • Identified that planning was skipped due to time pressure, not lack of ability.

Tasks

Draft a one-page breakdown of the next feature before writing any code.

Add a "Scope & Structure" section to PR descriptions for reviewers.

Flag to Alicia when a feature feels too large to plan in one sitting.

Created by HideMaru

from the Noun Project

Home

My Performance

My Team

Coaching

Performance Criterias

Administration

User Management

Performance

Team Management

Integrations

Alicia Walter

alicia@aude.ai

result

Engineers and managers appreciated having more context, more input, and a human in the room

I really appreciated the pattern cluster feature in the peer feedback view because it provides a more holistic view of actionable areas an engineer can focus on

— Engineering manager, usability testing

I really like the idea of keeping humans in the loop during coaching instead of relying solely on AI , I don’t fully trust AI on its own yet

— Engineer, usability testing

reflection

What I'd do differently

1

Design flexible coaching flows

Testing showed engineers come with different needs, some want 1:1s, others need structured resources, and coaching style differs per manager.

2

Explore better visualizations

The bubble chart made intuitive sense to some engineers but confused others. Frequency as a signal isn't always obvious at first glance.

What this project taught me

AI is only as useful as the trust behind it

Transparency isn't optional when AI touches something as sensitive as performance evaluation. Even with solid data sources, users need to understand the reasoning before they'll act on it.

Done is better than perfect

This project kept reminding me that trying to perfect something before putting it in front of users slows everything down. The best feedback came after something was in people's hands.

Designs need a path to production

I learned to keep one foot in what's actually buildable, not to limit ideas, but to make sure they have a real path to production.

More projects and experiences are on the way.
Let's keep in touch!

Email

Copied