Automating Insight: Saving Teams 6 Hours a Week with AI-Powered Bot Analysis

“We had dashboards, but we still didn’t know why the bot was failing.”

Brands invest heavily in automating support, but most teams are left with shallow metrics like handover rates or resolution times. That’s where Analyze comes in—an AI-powered analytics platform that surfaces topic-level insights, tracks sentiment, and learns over time to drive continuous improvement.

Note: I’ve used randomly generated data in the screenshots to protect client confidentiality. The visuals show real features, but the content is illustrative.

THE PROBLEM

Too much data, not enough direction — teams were stuck digging through chat logs without knowing what to fix.

My Role

8 people (PM, 2 designers, 4 devs, QA)

Team

Lead Product Designer

Timeline

4 months

Key CHALLENGES

Turning unstructured conversations
into clear insights

Challenge

Raw chatbot logs were all over the place—messy, inconsistent, and hard to use.

Solution

I added structure with Progressive Disclosure. It cut the noise, surfaced what mattered, and gave teams clarity without overload.

Designing on the go

Challenge

AI moves fast; models change, timelines shrink. The tricky part was keeping clarity while moving quickly.

Solution

I leaned on quick research, rough prototypes, and close teamwork with data and engineering. That way, we adapted fast without hurting usability.

MY DESIGN PROCESS

Build with users, not just for them

This wasn’t just a design sprint —
it was a continuous loop of insight → validation → iteration.

RESEARCH

From noise to clarity

Our goal was to understand not just what metrics users needed but how they made decisions about bot performance and improvements. Here's how we approached it:

User persona

We interviewed internal CS teams and reviewed user behaviors to build two detailed personas:

Customer success manager

Aisha Mehra

33 y/o, CSM

“I need to know where the bot is failing and how to fix it fast.”

🎯

Goals

Understand where the bot is failing and how to fix it fast.

😓

Frustrations

Relies on manual log reviews/spreadsheets, no clarity on whether issues stem from bot flow, content, or intent.

👇

Needs

Clear view of unresolved queries, sentiment trends, and actionable insights to improve containment.

🧠

How Might We

Summarize unresolved queries, link metrics to conversations, and surface automation opportunities to reduce prep time for QBRs.

⚙️

Behaviors & Tool Gaps

Tasks

Reviewing issues

Client reporting

Automation tracking

Current Behavior

Reads raw logs

Manual exports

Gut feeling

Limitations

Time-consuming

Not scalable

Risky

Bot Builder / Automation Specialist

“I need to know where the bot is failing and how to fix it fast.”

Aisha Mehra

33 y/o, CSM

🎯

Goals

Fix flows fast, cut unresolved queries, and boost CSAT with smoother self-service.

😓

Frustrations

Lacks a clear view of flow issues, switches tools often, and wastes time finding root causes.

👇

Needs

Quick filters, context-rich logs, and AI suggestions for KB and resolutions.

🧠

How Might We

Help Ravi find issues faster and fix them with less effort.

⚙️

Behaviors & Tool Gaps

Tasks

Fixing flows

Reviewing sentiment

Updating KB

Current Behavior

Manual debugging

Ignores it

Based on gut

Limitations

Slow & reactive

Not connected to flows

Not data-driven

Research Methods

Affinity mapping

Grouped feedback into six themes to uncover recurring pain points in analysis workflows.

Key inshights

  • Unclear resolution & sentiment

  • Disconnected dashboards

  • Manual reviews took 5–6 hours weekly

  • Agent replies weren’t reused

  • Data lacked “how to fix” context

Impact v/s effort analysis

We prioritized what to build based on the cost vs value of solving each pain point.

Competitive Audit

We evaluated how leading platforms like Ada and Intercom handled similar problems.

Targeted pain points

Reduce manual effort and increase decision clarity

After mapping feedback, identifying behavioural gaps, and evaluating competitor gaps, we narrowed down to the most critical problems our MVP needed to address. These were selected based on a combination of user pain, feasibility, and impact potential.

Manual effort to tag or categorize issues

CSMs and bot builders spent hours reviewing logs weekly. We saw teams averaging 5–6 hours/week just to prepare reports.

Auto tag conversations

Enable auto-tagging with LLMs, allowing users to filter by topic, failure, intent, or escalation

No clarity on resolution

Teams couldn’t tell if a user query was actually resolved, even if the bot completed a flow. This limited trust in the data.

Conversation classification

Give resolution status. Classify each conversation as resolved, unresolved, or handed over

No actionable suggestions

Even with charts and trends, teams didn’t know what to do next. Fixing issues relied heavily on intuition.

Actionable suggestions

Provide insight that highlight top failure reasons, drop-off patterns, and automation suggestions

No automation signals from agent replies

Valuable agent messages were not reused, even when they could have trained the bot. Automation potential was left untapped.

Auto learn from agent messages

Build an engine to identify repetitive agent replies that can be converted to KB articles or bot responses.

Disconnected dashboards and logs

Teams couldn't drill down from metrics to actual conversations. This broke the feedback loop and slowed issue resolution.

Add conversation to metrics

Link every metric to specific conversations, making insights traceable to the source

DISCOVERY

Validating hypothesized solution

Validation

Co-built with early customers by manually tagging chat logs to reveal why bots failed. Expanded across industries, refined tags, and introduced CRR & sentiment. Partnered with engineering to scale into an AI model that automated tagging and surfaced clear automation opportunities.

Client feedback

When we presented the insights, the reaction was immediate

“This is the first time we’ve seen why our bot is actually failing.”

“Can you make this process of analysis faster and smarter ”

The structured clarity clicked. For them and for us.

Key insights

  • Manual → AI tagging cut effort

  • Resolution tagging unlocked true success

  • Reports linked metrics to conversations

  • Agent replies revealed automation gaps

  • Need for filters & cleaner handover

  • Insights 3× faster on average

AI MODEL

Reduce manual effort and increase decision clarity

To move beyond dashboards, we started thinking differently, what if the bot could learn directly from real failures?

Our engineering team built an LLM-based system that analyzes conversations between agents and customers to surface where automation failed and how it could improve.

Conversations

AI model reviews conversations that were transferred to a human and got resolved

An AI model reviews conversations

Labels each conversation

Analyse how human solved query

Suggests articles to add to bot knowledge to improve automation

Topic

Model assigns a topic based on user intent

Containment

Checks if the query was resolved with/without agent handover

Resolution status

Checks if the query was resolved or not

User sentiment

Detects final user sentiment — positive, neutral, or negative

Design

Turning vision into reatilty

Analyze is built for fast-moving teams who can’t waste time in chat logs or static dashboards — it turns conversations into clear insights so they know exactly what to fix and why.

Topics list

We turned noisy chat data into a clear table that helps teams spot trends, gaps, and next steps.

  • Familiar table view, enriched with AI metrics like sentiment and resolution rate.

  • Every row highlights an action showing what’s working and what needs fixing.

  • Clear, high signal insights reduce overload and give teams confidence to act.

Topic details

Users needed more than high-level stats they wanted to see why a topic was struggling and what to do about it. This screen made that possible.

  • Key metrics up front show how often a topic appears, how bots handle it, and how users feel.

  • Clear next steps guide teams to act fast, like adding an article or resolving chats.

  • Smart suggestions explain the “why,” giving teams confidence to improve without analysts.

Conversation analysis

We connected insights to real conversations so teams could see exactly where the bot failed and why. By pairing full transcripts with AI analysis, users got both the story and the signal in one place.

  • Searchable queries, full transcripts, and AI tags are all shown side by side.

  • Unresolved chats are flagged instantly, making failures easy to spot.

  • Raw conversations build trust, while reducing screen-hopping speeds up diagnosis and decisions.

Outcome

Assessing the impact

We rolled out in phases, first to 15 enterprise customers for beta, then to our entire base. Every feedback loop helped us refine the product before the big launch.

30%

faster analysis saving teams 5–6 hours weekly

3x

quicker issue detection with AI driven tagging and insights

10%

increase in CRR for top accounts during beta

200+

users subscribed to analyse module

Improvements and future

While the early impact was promising, real user behavior uncovered areas for growth. Many teams weren’t acting on KB suggestions a sign we needed better integration with content workflows. We also heard a clear ask for a holistic view of topic performance, not just isolated metrics. Moving forward, we’re building features like multi-language support, deeper reporting, and easier adoption paths to turn insights into real action faster, and for more teams.

Thanks for reading! :)

Automating Insight: Saving Teams 6 Hours a Week with AI-Powered Bot Analysis

“We had dashboards, but we still didn’t know why the bot was failing.”

Brands invest heavily in automating support, but most teams are left with shallow metrics like handover rates or resolution times. That’s where Analyze comes in—an AI-powered analytics platform that surfaces topic-level insights, tracks sentiment, and learns over time to drive continuous improvement.

Note: I’ve used randomly generated data in the screenshots to protect client confidentiality. The visuals show real features, but the content is illustrative.

THE PROBLEM

Too much data, not enough direction — teams were stuck digging through chat logs without knowing what to fix.

My Role

8 people (PM, 2 designers, 4 devs, QA)

Team

Lead Product Designer

Timeline

4 months

Key CHALLENGES

Turning unstructured conversations
into clear insights

Challenge

Raw chatbot logs were all over the place—messy, inconsistent, and hard to use.

Solution

I added structure with Progressive Disclosure. It cut the noise, surfaced what mattered, and gave teams clarity without overload.

Designing on the go

Challenge

AI moves fast; models change, timelines shrink. The tricky part was keeping clarity while moving quickly.

Solution

I leaned on quick research, rough prototypes, and close teamwork with data and engineering. That way, we adapted fast without hurting usability.

MY DESIGN PROCESS

Build with users, not just for them

This wasn’t just a design sprint —
it was a continuous loop of insight → validation → iteration.

RESEARCH

From noise to clarity

Our goal was to understand not just what metrics users needed but how they made decisions about bot performance and improvements. Here's how we approached it:

User persona

We interviewed internal CS teams and reviewed user behaviors to build two detailed personas:

Customer success manager

Aisha Mehra

33 y/o, CSM

“I need to know where the bot is failing and how to fix it fast.”

🎯

Goals

Understand where the bot is failing and how to fix it fast.

😓

Frustrations

Relies on manual log reviews/spreadsheets, no clarity on whether issues stem from bot flow, content, or intent.

👇

Needs

Clear view of unresolved queries, sentiment trends, and actionable insights to improve containment.

🧠

How Might We

Summarize unresolved queries, link metrics to conversations, and surface automation opportunities to reduce prep time for QBRs.

⚙️

Behaviors & Tool Gaps

Tasks

Reviewing issues

Client reporting

Automation tracking

Current Behavior

Reads raw logs

Manual exports

Gut feeling

Limitations

Time-consuming

Not scalable

Risky

Bot Builder / Automation Specialist

“I need to know where the bot is failing and how to fix it fast.”

Aisha Mehra

33 y/o, CSM

🎯

Goals

Fix flows fast, cut unresolved queries, and boost CSAT with smoother self-service.

😓

Frustrations

Lacks a clear view of flow issues, switches tools often, and wastes time finding root causes.

👇

Needs

Quick filters, context-rich logs, and AI suggestions for KB and resolutions.

🧠

How Might We

Help Ravi find issues faster and fix them with less effort.

⚙️

Behaviors & Tool Gaps

Tasks

Fixing flows

Reviewing sentiment

Updating KB

Current Behavior

Manual debugging

Ignores it

Based on gut

Limitations

Slow & reactive

Not connected to flows

Not data-driven

Research Methods

Affinity mapping

Grouped feedback into six themes to uncover recurring pain points in analysis workflows.

Key inshights

  • Unclear resolution & sentiment

  • Disconnected dashboards

  • Manual reviews took 5–6 hours weekly

  • Agent replies weren’t reused

  • Data lacked “how to fix” context

Impact v/s effort analysis

We prioritized what to build based on the cost vs value of solving each pain point.

Competitive Audit

We evaluated how leading platforms like Ada and Intercom handled similar problems.

Targeted pain points

Reduce manual effort and increase decision clarity

After mapping feedback, identifying behavioural gaps, and evaluating competitor gaps, we narrowed down to the most critical problems our MVP needed to address. These were selected based on a combination of user pain, feasibility, and impact potential.

Manual effort to tag or categorize issues

CSMs and bot builders spent hours reviewing logs weekly. We saw teams averaging 5–6 hours/week just to prepare reports.

Auto tag conversations

Enable auto-tagging with LLMs, allowing users to filter by topic, failure, intent, or escalation

No clarity on resolution

Teams couldn’t tell if a user query was actually resolved, even if the bot completed a flow. This limited trust in the data.

Conversation classification

Give resolution status. Classify each conversation as resolved, unresolved, or handed over

No actionable suggestions

Even with charts and trends, teams didn’t know what to do next. Fixing issues relied heavily on intuition.

Actionable suggestions

Provide insight that highlight top failure reasons, drop-off patterns, and automation suggestions

No automation signals from agent replies

Valuable agent messages were not reused, even when they could have trained the bot. Automation potential was left untapped.

Auto learn from agent messages

Build an engine to identify repetitive agent replies that can be converted to KB articles or bot responses.

Disconnected dashboards and logs

Teams couldn't drill down from metrics to actual conversations. This broke the feedback loop and slowed issue resolution.

Add conversation to metrics

Link every metric to specific conversations, making insights traceable to the source

DISCOVERY

Validating hypothesized solution

Validation

Co-built with early customers by manually tagging chat logs to reveal why bots failed. Expanded across industries, refined tags, and introduced CRR & sentiment. Partnered with engineering to scale into an AI model that automated tagging and surfaced clear automation opportunities.

Client feedback

When we presented the insights, the reaction was immediate

“This is the first time we’ve seen why our bot is actually failing.”

“Can you make this process of analysis faster and smarter ”

The structured clarity clicked. For them and for us.

Key insights

  • Manual → AI tagging cut effort

  • Resolution tagging unlocked true success

  • Reports linked metrics to conversations

  • Agent replies revealed automation gaps

  • Need for filters & cleaner handover

  • Insights 3× faster on average

AI MODEL

Reduce manual effort and increase decision clarity

To move beyond dashboards, we started thinking differently, what if the bot could learn directly from real failures?

Our engineering team built an LLM-based system that analyzes conversations between agents and customers to surface where automation failed and how it could improve.

Conversations

AI model reviews conversations that were transferred to a human and got resolved

An AI model reviews conversations

Labels each conversation

Analyse how human solved query

Suggests articles to add to bot knowledge to improve automation

Topic

Model assigns a topic based on user intent

Containment

Checks if the query was resolved with/without agent handover

Resolution status

Checks if the query was resolved or not

User sentiment

Detects final user sentiment — positive, neutral, or negative

Design

Turning vision into reatilty

Analyze is built for fast-moving teams who can’t waste time in chat logs or static dashboards — it turns conversations into clear insights so they know exactly what to fix and why.

Topics list

We turned noisy chat data into a clear table that helps teams spot trends, gaps, and next steps.

  • Familiar table view, enriched with AI metrics like sentiment and resolution rate.

  • Every row highlights an action showing what’s working and what needs fixing.

  • Clear, high signal insights reduce overload and give teams confidence to act.

Topic details

Users needed more than high-level stats they wanted to see why a topic was struggling and what to do about it. This screen made that possible.

  • Key metrics up front show how often a topic appears, how bots handle it, and how users feel.

  • Clear next steps guide teams to act fast, like adding an article or resolving chats.

  • Smart suggestions explain the “why,” giving teams confidence to improve without analysts.

Conversation analysis

We connected insights to real conversations so teams could see exactly where the bot failed and why. By pairing full transcripts with AI analysis, users got both the story and the signal in one place.

  • Searchable queries, full transcripts, and AI tags are all shown side by side.

  • Unresolved chats are flagged instantly, making failures easy to spot.

  • Raw conversations build trust, while reducing screen-hopping speeds up diagnosis and decisions.

Outcome

Assessing the impact

We rolled out in phases, first to 15 enterprise customers for beta, then to our entire base. Every feedback loop helped us refine the product before the big launch.

30%

faster analysis saving teams 5–6 hours weekly

3x

quicker issue detection with AI driven tagging and insights

10%

increase in CRR for top accounts during beta

200+

users subscribed to analyse module

Improvements and future

While the early impact was promising, real user behavior uncovered areas for growth. Many teams weren’t acting on KB suggestions a sign we needed better integration with content workflows. We also heard a clear ask for a holistic view of topic performance, not just isolated metrics. Moving forward, we’re building features like multi-language support, deeper reporting, and easier adoption paths to turn insights into real action faster, and for more teams.

Thanks for reading! :)

Automating Insight: Saving Teams 6 Hours a Week with AI-Powered Bot Analysis

“We had dashboards, but we still didn’t know why the bot was failing.”

Brands invest heavily in automating support, but most teams are left with shallow metrics like handover rates or resolution times. That’s where Analyze comes in—an AI-powered analytics platform that surfaces topic-level insights, tracks sentiment, and learns over time to drive continuous improvement.

Note: I’ve used randomly generated data in the screenshots to protect client confidentiality. The visuals show real features, but the content is illustrative.

THE PROBLEM

Too much data, not enough direction — teams were stuck digging through chat logs without knowing what to fix.

My Role

8 people (PM, 2 designers, 4 devs, QA)

Team

Lead Product Designer

Timeline

4 months

Key CHALLENGES

Turning unstructured conversations
into clear insights

Challenge

Raw chatbot logs were all over the place—messy, inconsistent, and hard to use.

Solution

I added structure with Progressive Disclosure. It cut the noise, surfaced what mattered, and gave teams clarity without overload.

Designing on the go

Challenge

AI moves fast; models change, timelines shrink. The tricky part was keeping clarity while moving quickly.

Solution

I leaned on quick research, rough prototypes, and close teamwork with data and engineering. That way, we adapted fast without hurting usability.

MY DESIGN PROCESS

Build with users, not just for them

This wasn’t just a design sprint —
it was a continuous loop of insight → validation → iteration.

RESEARCH

From noise to clarity

Our goal was to understand not just what metrics users needed but how they made decisions about bot performance and improvements. Here's how we approached it:

User persona

We interviewed internal CS teams and reviewed user behaviors to build two detailed personas:

Customer success manager

Aisha Mehra

33 y/o, CSM

“I need to know where the bot is failing and how to fix it fast.”

🎯

Goals

Understand where the bot is failing and how to fix it fast.

😓

Frustrations

Relies on manual log reviews/spreadsheets, no clarity on whether issues stem from bot flow, content, or intent.

👇

Needs

Clear view of unresolved queries, sentiment trends, and actionable insights to improve containment.

🧠

How Might We

Summarize unresolved queries, link metrics to conversations, and surface automation opportunities to reduce prep time for QBRs.

⚙️

Behaviors & Tool Gaps

Tasks

Reviewing issues

Client reporting

Automation tracking

Current Behavior

Reads raw logs

Manual exports

Gut feeling

Limitations

Time-consuming

Not scalable

Risky

Bot Builder / Automation Specialist

“I need to know where the bot is failing and how to fix it fast.”

Aisha Mehra

33 y/o, CSM

🎯

Goals

Fix flows fast, cut unresolved queries, and boost CSAT with smoother self-service.

😓

Frustrations

Lacks a clear view of flow issues, switches tools often, and wastes time finding root causes.

👇

Needs

Quick filters, context-rich logs, and AI suggestions for KB and resolutions.

🧠

How Might We

Help Ravi find issues faster and fix them with less effort.

⚙️

Behaviors & Tool Gaps

Tasks

Fixing flows

Reviewing sentiment

Updating KB

Current Behavior

Manual debugging

Ignores it

Based on gut

Limitations

Slow & reactive

Not connected to flows

Not data-driven

Research Methods

Affinity mapping

Grouped feedback into six themes to uncover recurring pain points in analysis workflows.

Key inshights

  • Unclear resolution & sentiment

  • Disconnected dashboards

  • Manual reviews took 5–6 hours weekly

  • Agent replies weren’t reused

  • Data lacked “how to fix” context

Impact v/s effort analysis

We prioritized what to build based on the cost vs value of solving each pain point.

Competitive Audit

We evaluated how leading platforms like Ada and Intercom handled similar problems.

Targeted pain points

Reduce manual effort and increase decision clarity

After mapping feedback, identifying behavioural gaps, and evaluating competitor gaps, we narrowed down to the most critical problems our MVP needed to address. These were selected based on a combination of user pain, feasibility, and impact potential.

Manual effort to tag or categorize issues

CSMs and bot builders spent hours reviewing logs weekly. We saw teams averaging 5–6 hours/week just to prepare reports.

Auto tag conversations

Enable auto-tagging with LLMs, allowing users to filter by topic, failure, intent, or escalation

No clarity on resolution

Teams couldn’t tell if a user query was actually resolved, even if the bot completed a flow. This limited trust in the data.

Conversation classification

Give resolution status. Classify each conversation as resolved, unresolved, or handed over

No actionable suggestions

Even with charts and trends, teams didn’t know what to do next. Fixing issues relied heavily on intuition.

Actionable suggestions

Provide insight that highlight top failure reasons, drop-off patterns, and automation suggestions

No automation signals from agent replies

Valuable agent messages were not reused, even when they could have trained the bot. Automation potential was left untapped.

Auto learn from agent messages

Build an engine to identify repetitive agent replies that can be converted to KB articles or bot responses.

Disconnected dashboards and logs

Teams couldn't drill down from metrics to actual conversations. This broke the feedback loop and slowed issue resolution.

Add conversation to metrics

Link every metric to specific conversations, making insights traceable to the source

DISCOVERY

Validating hypothesized solution

Validation

Co-built with early customers by manually tagging chat logs to reveal why bots failed. Expanded across industries, refined tags, and introduced CRR & sentiment. Partnered with engineering to scale into an AI model that automated tagging and surfaced clear automation opportunities.

Client feedback

When we presented the insights, the reaction was immediate

“This is the first time we’ve seen why our bot is actually failing.”

“Can you make this process of analysis faster and smarter ”

The structured clarity clicked. For them and for us.

Key insights

  • Manual → AI tagging cut effort

  • Resolution tagging unlocked true success

  • Reports linked metrics to conversations

  • Agent replies revealed automation gaps

  • Need for filters & cleaner handover

  • Insights 3× faster on average

AI MODEL

Reduce manual effort and increase decision clarity

To move beyond dashboards, we started thinking differently, what if the bot could learn directly from real failures?

Our engineering team built an LLM-based system that analyzes conversations between agents and customers to surface where automation failed and how it could improve.

Conversations

AI model reviews conversations that were transferred to a human and got resolved

An AI model reviews conversations

Labels each conversation

Analyse how human solved query

Suggests articles to add to bot knowledge to improve automation

Topic

Model assigns a topic based on user intent

Containment

Checks if the query was resolved with/without agent handover

Resolution status

Checks if the query was resolved or not

User sentiment

Detects final user sentiment — positive, neutral, or negative

Design

Turning vision into reatilty

Analyze is built for fast-moving teams who can’t waste time in chat logs or static dashboards — it turns conversations into clear insights so they know exactly what to fix and why.

Topics list

We turned noisy chat data into a clear table that helps teams spot trends, gaps, and next steps.

  • Familiar table view, enriched with AI metrics like sentiment and resolution rate.

  • Every row highlights an action showing what’s working and what needs fixing.

  • Clear, high signal insights reduce overload and give teams confidence to act.

Topic details

Users needed more than high-level stats they wanted to see why a topic was struggling and what to do about it. This screen made that possible.

  • Key metrics up front show how often a topic appears, how bots handle it, and how users feel.

  • Clear next steps guide teams to act fast, like adding an article or resolving chats.

  • Smart suggestions explain the “why,” giving teams confidence to improve without analysts.

Conversation analysis

We connected insights to real conversations so teams could see exactly where the bot failed and why. By pairing full transcripts with AI analysis, users got both the story and the signal in one place.

  • Searchable queries, full transcripts, and AI tags are all shown side by side.

  • Unresolved chats are flagged instantly, making failures easy to spot.

  • Raw conversations build trust, while reducing screen-hopping speeds up diagnosis and decisions.

Outcome

Assessing the impact

We rolled out in phases, first to 15 enterprise customers for beta, then to our entire base. Every feedback loop helped us refine the product before the big launch.

30%

faster analysis saving teams 5–6 hours weekly

3x

quicker issue detection with AI driven tagging and insights

10%

increase in CRR for top accounts during beta

200+

users subscribed to analyse module

Improvements and future

While the early impact was promising, real user behavior uncovered areas for growth. Many teams weren’t acting on KB suggestions a sign we needed better integration with content workflows. We also heard a clear ask for a holistic view of topic performance, not just isolated metrics. Moving forward, we’re building features like multi-language support, deeper reporting, and easier adoption paths to turn insights into real action faster, and for more teams.

Thanks for reading! :)

Automating Insight: Saving Teams 6 Hours a Week with AI-Powered Bot Analysis

“We had dashboards, but we still didn’t know why the bot was failing.”

Brands invest heavily in automating support, but most teams are left with shallow metrics like handover rates or resolution times. That’s where Analyze comes in—an AI-powered analytics platform that surfaces topic-level insights, tracks sentiment, and learns over time to drive continuous improvement.

Note: I’ve used randomly generated data in the screenshots to protect client confidentiality. The visuals show real features, but the content is illustrative.

THE PROBLEM

Too much data, not enough direction — teams were stuck digging through chat logs without knowing what to fix.

My Role

8 people (PM, 2 designers, 4 devs, QA)

Team

Lead Product Designer

Timeline

4 months

Key CHALLENGES

Turning unstructured conversations
into clear insights

Challenge

Raw chatbot logs were all over the place—messy, inconsistent, and hard to use.

Solution

I added structure with Progressive Disclosure. It cut the noise, surfaced what mattered, and gave teams clarity without overload.

Designing on the go

Challenge

AI moves fast; models change, timelines shrink. The tricky part was keeping clarity while moving quickly.

Solution

I leaned on quick research, rough prototypes, and close teamwork with data and engineering. That way, we adapted fast without hurting usability.

MY DESIGN PROCESS

Build with users, not just for them

This wasn’t just a design sprint —
it was a continuous loop of insight → validation → iteration.

RESEARCH

From noise to clarity

Our goal was to understand not just what metrics users needed but how they made decisions about bot performance and improvements. Here's how we approached it:

User persona

We interviewed internal CS teams and reviewed user behaviors to build two detailed personas:

Customer success manager

Aisha Mehra

33 y/o, CSM

“I need to know where the bot is failing and how to fix it fast.”

🎯

Goals

Understand where the bot is failing and how to fix it fast.

😓

Frustrations

Relies on manual log reviews/spreadsheets, no clarity on whether issues stem from bot flow, content, or intent.

👇

Needs

Clear view of unresolved queries, sentiment trends, and actionable insights to improve containment.

🧠

How Might We

Summarize unresolved queries, link metrics to conversations, and surface automation opportunities to reduce prep time for QBRs.

⚙️

Behaviors & Tool Gaps

Tasks

Reviewing issues

Client reporting

Automation tracking

Current Behavior

Reads raw logs

Manual exports

Gut feeling

Limitations

Time-consuming

Not scalable

Risky

Bot Builder / Automation Specialist

“I need to know where the bot is failing and how to fix it fast.”

Aisha Mehra

33 y/o, CSM

🎯

Goals

Fix flows fast, cut unresolved queries, and boost CSAT with smoother self-service.

😓

Frustrations

Lacks a clear view of flow issues, switches tools often, and wastes time finding root causes.

👇

Needs

Quick filters, context-rich logs, and AI suggestions for KB and resolutions.

🧠

How Might We

Help Ravi find issues faster and fix them with less effort.

⚙️

Behaviors & Tool Gaps

Tasks

Fixing flows

Reviewing sentiment

Updating KB

Current Behavior

Manual debugging

Ignores it

Based on gut

Limitations

Slow & reactive

Not connected to flows

Not data-driven

Research Methods

Affinity mapping

Grouped feedback into six themes to uncover recurring pain points in analysis workflows.

Key inshights

  • Unclear resolution & sentiment

  • Disconnected dashboards

  • Manual reviews took 5–6 hours weekly

  • Agent replies weren’t reused

  • Data lacked “how to fix” context

Impact v/s effort analysis

We prioritized what to build based on the cost vs value of solving each pain point.

Competitive Audit

We evaluated how leading platforms like Ada and Intercom handled similar problems.

Targeted pain points

Reduce manual effort and increase decision clarity

After mapping feedback, identifying behavioural gaps, and evaluating competitor gaps, we narrowed down to the most critical problems our MVP needed to address. These were selected based on a combination of user pain, feasibility, and impact potential.

Manual effort to tag or categorize issues

CSMs and bot builders spent hours reviewing logs weekly. We saw teams averaging 5–6 hours/week just to prepare reports.

Auto tag conversations

Enable auto-tagging with LLMs, allowing users to filter by topic, failure, intent, or escalation

No clarity on resolution

Teams couldn’t tell if a user query was actually resolved, even if the bot completed a flow. This limited trust in the data.

Conversation classification

Give resolution status. Classify each conversation as resolved, unresolved, or handed over

No actionable suggestions

Even with charts and trends, teams didn’t know what to do next. Fixing issues relied heavily on intuition.

Actionable suggestions

Provide insight that highlight top failure reasons, drop-off patterns, and automation suggestions

No automation signals from agent replies

Valuable agent messages were not reused, even when they could have trained the bot. Automation potential was left untapped.

Auto learn from agent messages

Build an engine to identify repetitive agent replies that can be converted to KB articles or bot responses.

Disconnected dashboards and logs

Teams couldn't drill down from metrics to actual conversations. This broke the feedback loop and slowed issue resolution.

Add conversation to metrics

Link every metric to specific conversations, making insights traceable to the source

DISCOVERY

Validating hypothesized solution

Validation

Co-built with early customers by manually tagging chat logs to reveal why bots failed. Expanded across industries, refined tags, and introduced CRR & sentiment. Partnered with engineering to scale into an AI model that automated tagging and surfaced clear automation opportunities.

Client feedback

When we presented the insights, the reaction was immediate

“This is the first time we’ve seen why our bot is actually failing.”

“Can you make this process of analysis faster and smarter ”

The structured clarity clicked. For them and for us.

Key insights

  • Manual → AI tagging cut effort

  • Resolution tagging unlocked true success

  • Reports linked metrics to conversations

  • Agent replies revealed automation gaps

  • Need for filters & cleaner handover

  • Insights 3× faster on average

AI MODEL

Reduce manual effort and increase decision clarity

To move beyond dashboards, we started thinking differently, what if the bot could learn directly from real failures?

Our engineering team built an LLM-based system that analyzes conversations between agents and customers to surface where automation failed and how it could improve.

Conversations

AI model reviews conversations that were transferred to a human and got resolved

An AI model reviews conversations

Labels each conversation

Analyse how human solved query

Suggests articles to add to bot knowledge to improve automation

Topic

Model assigns a topic based on user intent

Containment

Checks if the query was resolved with/without agent handover

Resolution status

Checks if the query was resolved or not

User sentiment

Detects final user sentiment — positive, neutral, or negative

Design

Turning vision into reatilty

Analyze is built for fast-moving teams who can’t waste time in chat logs or static dashboards — it turns conversations into clear insights so they know exactly what to fix and why.

Topics list

We turned noisy chat data into a clear table that helps teams spot trends, gaps, and next steps.

  • Familiar table view, enriched with AI metrics like sentiment and resolution rate.

  • Every row highlights an action showing what’s working and what needs fixing.

  • Clear, high signal insights reduce overload and give teams confidence to act.

Topic details

Users needed more than high-level stats they wanted to see why a topic was struggling and what to do about it. This screen made that possible.

  • Key metrics up front show how often a topic appears, how bots handle it, and how users feel.

  • Clear next steps guide teams to act fast, like adding an article or resolving chats.

  • Smart suggestions explain the “why,” giving teams confidence to improve without analysts.

Conversation analysis

We connected insights to real conversations so teams could see exactly where the bot failed and why. By pairing full transcripts with AI analysis, users got both the story and the signal in one place.

  • Searchable queries, full transcripts, and AI tags are all shown side by side.

  • Unresolved chats are flagged instantly, making failures easy to spot.

  • Raw conversations build trust, while reducing screen-hopping speeds up diagnosis and decisions.

Outcome

Assessing the impact

We rolled out in phases, first to 15 enterprise customers for beta, then to our entire base. Every feedback loop helped us refine the product before the big launch.

30%

faster analysis saving teams 5–6 hours weekly

3x

quicker issue detection with AI driven tagging and insights

10%

increase in CRR for top accounts during beta

200+

users subscribed to analyse module

Improvements and future

While the early impact was promising, real user behavior uncovered areas for growth. Many teams weren’t acting on KB suggestions a sign we needed better integration with content workflows. We also heard a clear ask for a holistic view of topic performance, not just isolated metrics. Moving forward, we’re building features like multi-language support, deeper reporting, and easier adoption paths to turn insights into real action faster, and for more teams.

Thanks for reading! :)

Got something cool in mind?

Coffee chats, collabs, or cosmic ideas—bring it on.

Got something cool in mind?

Coffee chats, collabs, or cosmic ideas—bring it on.

Got something cool in mind?

Coffee chats, collabs, or cosmic ideas—bring it on.

Got something cool in mind?

Coffee chats, collabs, or cosmic ideas—bring it on.

Play to listen

4:31 min.

Play to listen

4:31 min.

Create a free website with Framer, the website builder loved by startups, designers and agencies.