Performance reports

Overview

Measure your AI Agent’s performance with a variety of detailed reports. Reports help you understand conversation outcomes, identify trends, and make data-driven decisions to improve your AI Agent.

By default, reports do not include data from test users, so testing your AI Agent does not skew your results. Unless otherwise noted, you can view reports by going to Analytics > Reports in the Ada dashboard.

Use cases

Reports help you understand your AI Agent’s effectiveness and identify opportunities for improvement.

  • Track automated resolution: Monitor how often your AI Agent resolves inquiries without human intervention.
  • Identify knowledge gaps: Use Knowledge usage reports to find articles with low resolution rates that may need updates.
  • Monitor customer satisfaction: Review CSAT scores and survey results to understand end user sentiment.
  • Analyze Action performance: Track API usage, error rates, and resolution rates for each Action.
  • Understand conversation patterns: Review volume, handle time, and escalation trends over time.

Capabilities & configuration

Reports provide flexible options for analyzing and exporting your AI Agent’s performance data.

  • Date range filtering: View data for predefined periods or custom date ranges.
  • Multi-criteria filtering: Narrow results by AR classification, CSAT, channel, language, Actions, Playbooks, Topics, and more.
  • Conversation drill-through: Click metrics to view filtered conversations in the Convos view.
  • Export and print: Save reports as PDF or print them for offline review.
  • Test user exclusion: By default, test conversations are excluded from reports to ensure accurate metrics.

Report details

Automated resolution and containment

The automated resolution rate is an analysis of how many conversations your AI Agent was able to resolve automatically.

To calculate the automated resolution rate, your AI Agent analyzes each completed conversation to understand both the end user’s intent and the AI Agent’s response. Based on that analysis, it then assigns a classification of either Resolved or Not Resolved to each conversation.

For a conversation to be considered automatically resolved, the conversation must be:

  • Relevant - Ada effectively understood the end user’s inquiry, and provided directly related information or assistance.

  • Accurate - Ada provided correct, up-to-date information.

  • Safe - Ada interacted with the end user in a respectful manner and avoided engaging in topics that caused danger or harm.

  • Contained - Ada addressed the end user’s inquiry without having to hand them off to a human agent.

    While Containment Rate can be a useful metric to get a quick glance of the proportion of AI Agent conversations that did not escalate to a human agent, automated resolution rate takes it a step further. By measuring the success of those conversations and the content they contain, you can get a much better idea of how helpful your AI Agent really is.

Your AI Agent will only assess for automated resolution when a conversation has ended. When viewing the automated resolution rate graph, a dotted line may appear to indicate that recent conversations may not have ended and therefore may cause the automated resolution rate to fluctuate once they’re analyzed. For more information on how the conversation lifecycle impacts automated resolution, see automated resolution rate.

In this list, you can view a summary of what each end user was looking for, how your AI Agent classified the conversation, and its reasoning. If you need more information, you can click a row to view the entire conversation transcript.

MetricDefinition

Automated Resolution Rate

The percentage of conversations that your AI Agent determined were automatically resolved. Your AI Agent calculates this with the formula Resolved conversations / (Resolved conversations + Not Resolved conversations).

Containment Rate

The percent of conversations that did not result in a handoff to a human agent.

API usage

Provides visibility into how often Ada is performing each action, and highlights errors with full log download functionality - allowing your team to troubleshoot effectively. You can access this report through the Reports tab (under Performance) in the left navigation menu or directly through the report icon at the top of the Actions Hub.

MetricDefinition

Conversations

The number or percentage of conversations where a specific Action was used. Click on this number to see these conversations filtered in the Convos view.

API calls

The total number of API calls made by an Action.

Error Rate

The percentage of failed API calls made.

AR rate

The percentage of conversations that your AI Agent determined were automatically resolved. Your AI Agent calculates this with the formula Resolved conversations / (Resolved conversations + Not Resolved conversations).

Containment rate

The percent of conversations that did not result in a handoff to human support.

CSAT

The percent of conversations end users reviewed positively, out of all conversations they reviewed.

Agent satisfaction score

View customer satisfaction (CSAT) surveys where the scores are attributed to human support, available if the β€œAutomatically survey after chat” option is turned on.

When you filter this report by date, it uses the date that the user submitted their satisfaction survey, rather than the date the conversation started. As a result, the number of conversations that appear in this report may vary from other reports.

There are four ways you can set up customer satisfaction reviews, each with different scales for recording feedback:

Rating typeNegative reviewPositive review
Numeric (5-point scale)1, 2, or 34 or 5
Numeric (10-point scale)1, 2, 3, 4, 5, or 67, 8, 9, or 10
Emoji (5-point scale)😠, πŸ™, or πŸ˜πŸ™‚ or 😍
Thumbs up/down (binary)πŸ‘ŽπŸ‘
MetricDefinition

Live chat score

The percent of agent reviews that were positive. Your AI Agent calculates this with the formula SUM (positive agent reviews) / SUM (all agent reviews) * 100.

Agent name

The name of the agent who spoke with the end user immediately before the end user provided the review. If multiple agents interacted with the end user in the same conversation, even if only one agent’s name appears in this list, all of the agents in that conversation are assigned the end user’s CSAT score.

Agent names appear in this list if they have at least one review in the time periods selected for either data display or for comparison.

Avg score

The percent of agent reviews that were positive.

# of positive

The number of agent reviews that were positive.

# of negative

The number of agent reviews that were negative.

Total # of surveys

The total number of agent reviews.

Average handle time

View the average amount of time end users spent talking with your AI Agent, for conversations that did not end in handoffs to human support.

This report uses winsorization on all of its metrics. To handle outliers, your AI Agent calculates the 90th percentile of all handle times. If a handle time is higher than the 90th percentile limit, your AI Agent replaces it with the 90th percentile limit instead.

MetricDefinition
Avg handle time when containedThe average amount of time end users spent talking with your AI Agent, for conversations that did not end in handoffs to human support.
Avg handle time before escalationThe average amount of time end users spent talking to your AI Agent before handoff, for conversations where end users escalated to human support.
Avg handle time with agentsThe average amount of time end users spent talking to live support agents.

Conversational messages volume

View the number of AI Agent, end user, and human agent messages per conversation.

Example conversation:

AI Agent
- Hello! (1)
- Hello! How can I be of assistance today? (2)
End user
[1] Hello -
[2] What is the status of my order? -
AI Agent
- I can check on that for you. (3)
- What is your order number? (4)
End user
[3] abc123 -
AI Agent
- Let me fetch that information for you... (5)
- Your order is currently being packaged for shipping. (6)
- Your estimated delivery date is Dec 25. (7)
End user
[4] that is too long. let me speak to an agent -
AI Agent
- Understood. Connecting you to the next available agent (8)
Human agent
- Hello my name is Sonia. How can I further help you? {1}
End user
[5] I need my order sooner. please cancel it -
Human agent
- Sorry about the delay. I will cancel your order {2}
- Your order has been cancelled {3}
End user
[6] Thank you -
MetricDefinition

Number of conversations

The number of conversations where an end user sent at least one message to your AI Agent.

Messages sent

The number of conversations (y-axis) that contained a given number of messages your AI Agent sent (x-axis).

In the example above, where AI Agent messages are counted in parentheses (), this conversation would fall under 8 AI Agent messages. Each response bubble counts as a single message, excluding messages that indicate a live agent has joined or left the chat.

Customer messages received

The number of conversations (y-axis) that contained a given number of messages customers sent (x-axis).

In the example above, where customer messages are counted in square brackets [], this conversation would fall under 6 customer messages.

Agent messages

The number of conversations (y-axis) that contained a given number of messages agents sent (x-axis).

In the example above, where agent messages are counted in curly brackets {}, this conversation would fall under 3 agent messages. Emojis, links, and pictures all count as agent messages for this report.

Number of messages (x-axis)

The number of each type of message per conversation.

Roughly 95% of conversations have fewer than 45 messages of any one type, which is why the upper end of the scale groups all conversations with 45 or more of any one type of message.

Number of conversations (y-axis)

The number of conversations that fall in each quantity of messages.

Conversations breakdown

View the number of conversations initiated, engaged, and escalated in your AI Agent.

MetricDefinition

Opened

The number of conversations where an end user initiated contact and was presented with a greeting. Every conversation contains one greeting. The entire series of messages that may be sent counts as one greeting, but only one needs to be sent for it to count as an open.

Engaged

The number of conversations where an end user sent at least one message or interacted at least once.

A conversation counts as engaged once an end user sends a message, regardless of whether your AI Agent understands the message.

Escalated

The number of conversations that resulted in a handoff to human support.

Autonomously Resolved

The number of conversations that your AI Agent contained and resolved.

Before July 31, 2024, this number was approximated based on the automated resolution rate (AR%) of a sample of your conversations, and was calculated with the formula # of engaged conversations x AR%.

The calculated number of automatically resolved conversations was subject to the error margin of the calculated AR%.

For more information, see Understand and improve your AI Agent’s automated resolution rate.

Customer satisfaction score

View the percent of your AI Agent’s conversations that end users reviewed positively. For more information, see Collect and analyze customer satisfaction data with Satisfaction Surveys.

There are four ways you can set up customer satisfaction reviews, each with different scales for recording feedback:

Rating typeNegative reviewPositive review
Numeric (5-point scale)1, 2, or 34 or 5
Numeric (10-point scale)1, 2, 3, 4, 5, or 67, 8, 9, or 10
Emoji (5-point scale)😠, πŸ™, or πŸ˜πŸ™‚ or 😍
Thumbs up/down (binary)πŸ‘ŽπŸ‘
MetricDefinition
Overall scoreThe percent of conversations end users reviewed positively, out of all conversations they reviewed.

Knowledge usage

View to help you understand which articles are most frequently used by Ada in end user responses, and which articles are correlated with high or low Automated Resolution Rates as well as other performance metrics. Includes conversation drill-throughs to support improvement workflows. You can access this report through the Reports tab (under Performance) in the left navigation menu or directly through the report icon at the top of the Knowledge Hub.

MetricDefinition

Conversations

The number or percentage of conversations where a specific article was used. Click on this number to see these conversations filtered in the Convos view.

AR rate

The percentage of conversations that your AI Agent determined were automatically resolved. Your AI Agent calculates this with the formula Resolved conversations / (Resolved conversations + Not Resolved conversations).

Containment rate

The percent of conversations that did not result in a handoff to human support.

CSAT

The percent of conversations end users reviewed positively, out of all conversations they reviewed.

Satisfaction survey results

View the results of your customer satisfaction (CSAT) survey. For more information, see Collect and analyze customer satisfaction data with Satisfaction Surveys.

When you filter this report by date, it uses the date that the user submitted their satisfaction survey, rather than the date the conversation started. As a result, the number of conversations that appear in this report may vary from other reports.

There are four ways you can set up customer satisfaction reviews, each with different scales for recording feedback:

Rating typeNegative reviewPositive review
Numeric (5-point scale)1, 2, or 34 or 5
Numeric (10-point scale)1, 2, 3, 4, 5, or 67, 8, 9, or 10
Emoji (5-point scale)😠, πŸ™, or πŸ˜πŸ™‚ or 😍
Thumbs up/down (binary)πŸ‘ŽπŸ‘
MetricDefinition

Last submitted

The most recent time an end user submitted a satisfaction survey.

Agent

The agent, if any, who participated in the conversation. If multiple agents participated in the conversation, this is the agent who participated closest to the end of the chat.

Survey type

The type of survey the end user responded to.

  • End chat: The survey presented to the end user when they click β€œEnd chat” outside of a handoff.

  • Live agent: The survey end users receive when they close the chat after speaking with an agent, or when an agent leaves the conversation.

CSAT

The satisfaction rating the end user selected.

Reason for rating

The reason(s) that the end user selected in the survey follow-up question, if any.

Possible positive reasons:

  • Efficient chat

  • Helpful resolution

  • Knowledgeable support

  • Friendly tone

  • Easy to use

  • Other

Possible negative reasons:

  • Took too long

  • Unhelpful resolution

  • Lack of expertise

  • Unfriendly tone

  • Technical issues

  • Other

Resolution

The end user’s response, if any, to whether your AI Agent was able to resolve their issue. This can either be yes or no.

Customer Effort Score

The end user’s effort score rating indicating how easy it was to get help. Uses a 5-point or 7-point scale.

Net Promoter Score

The end user’s likelihood to recommend rating on a 0-10 scale. Scores 0-6 are Detractors, 7-8 are Passives, and 9-10 are Promoters.

Comments

Additional comments, if any, that the end user wanted to include in the survey about their experience.

Proactive conversations

View a detailed breakdown of how your AI Agent is engaging end users through Proactive conversations and how effectively those interactions contribute to automated resolutions and customer satisfaction.

At the top of the report, you’ll see a graph that compares:

  • All Conversations: The total number of conversations that occurred within the selected date range.
  • Proactive Conversations: The number of conversations initiated by your AI Agent through Proactive messages that received at least one end user response.

This graph allows you to understand the reach and uptake of Proactive messaging in the context of your broader end user engagement volume. By comparing trends in Proactive conversations to overall volume, you can assess how actively end users are engaging with proactive outreach efforts and identify opportunities to refine your messaging strategy.

Beneath the graph, a table gives you a detailed breakdown of how each individual Proactive conversation is performing. Each row represents a specific Proactive conversation, with the following metrics displayed:

MetricDescription
ConversationsThe number of end user conversations initiated by this Proactive message over the selected period.
% of ConversationsThe percentage that this Proactive conversation represents out of the total conversation volume.
AR RateThe Automated Resolution rate β€” the percentage of conversations that were fully resolved by the AI Agent without requiring human support.
ContainmentThe percentage of conversations that used this Proactive message and were contained by the AI Agent β€” meaning no handoff to human support.
CSATThe Customer Satisfaction score β€” the percentage of conversations using this message that received a positive CSAT response, based only on those rated by end users.

Quick start

View and filter report data in a few steps.

1

On the Ada dashboard, go to Analytics > Reports.

2

Select a report from the list (for example, Automated resolution and containment).

3

Use the date filter to select a time range, then click Apply.

4

Optionally, click Add Filter to narrow results by criteria such as Channel, Language, or AR classification.

For detailed filtering options, see Implementation & usage.

Implementation & usage

Filter and export report data to focus on the metrics that matter most to your analysis.

Filter by date

To filter a report by date:

  1. Click the date filter drop-down.

  2. Define your date range by one of the following:

    • Select a predefined range from the list on the left.

    • Type the filter start date in the Starting field. Type the filter end date in the Ending field.

    • Click the starting date on the calendar on the left, and the ending date on the calendar on the right.

  3. Click Apply.

The date filter dropdown provides you with the ability to specify the date range you want to filter the report’s data by. You can select from a list of preset date ranges or select Custom… to specify your own by way of a calendar selector.

Filter by additional criteria

The list of available filters differs for each report, depending on the data the report includes. Clicking the Add Filter drop-down menu gives you access to the filters relevant to the report you’re viewing.

Use these options to control which data appears in a report.

  • Include test user: Include data from conversations originating from the Ada dashboard test AI Agent. Test conversations are excluded by default.

  • AR classification: The automatic resolution classification for the conversation.

  • Coaching: Conversations where one or more Coaching instructions were applied.

  • CSAT: Customer satisfaction (CSAT) ratings submitted by end users.

  • Article: Conversations that referenced one or more specific Knowledge articles.

  • Action: Conversations associated with one or more Actions.

  • Playbook: Conversations associated with one or more Playbooks.

  • Conversation category: Conversations whose assigned Topics have been grouped under one or more categories.

  • Generated topic: Conversations your AI Agent automatically assigned to one or more Topics.

  • Engaged: Conversations where an end user sent at least one message.

  • Handoff: Conversations that resulted in a Handoff to a human agent.

  • Language (Multilingual feature required): View reporting analytics that referenced conversations in one or more specific Languages.

  • Channel: The Channel where the conversation took place. For example, Ada Web Chat, SMS, WhatsApp, and so on.

  • Browser: Conversations where end users used specific browsers. For example, Chrome, Firefox, Safari, and so on.

  • Device: Conversations where end users used a specific device or operating system. For example, Windows, iPhone, Android, and so on.

  • Live agent: Conversations that involved one or more human agents.

  • Status code: Conversations that include API calls that resulted with one or more specific error code types. For example, 1xx, 2xx, 3xx, and so on.

  • Agent review: Conversations that include a human agent’s review.

  • Reason for rating: Conversations where end users selected one or more specific reasons when submitting a CSAT rating.

  • Variable: Conversations that include one or more Variables. You can filter by specific values or by whether a Variable Is Set or Is Not Set.

Additional information

  • Report data is updated approximately every hour (but may take up to three hours).

  • Reports are in the time zone set in your profile.

Print a report

For the best experience, view reports directly in the dashboard. If you need to save a report as a PDF or print it, use these settings to limit rendering issues.

To print or save a report as PDF:

  1. Click Print.

  2. In the Print window that appears, beside Destination, select either Save as PDF or a printer.

  3. Click More settings to display additional print settings.

  4. Set Margins to Minimum.

  5. Set Scale to Custom, then change the value to 70.

    • Alternatively, you can set the Paper size to A3 (11-3/4 x 16-1/2 in) or Legal (8.5 x 14 in).
  6. Under Options, select the Background graphics check box.

  7. Right before saving or printing, scroll through your print preview, and beside Pages, change the number of pages you want to include in your PDF or printout. The settings you changed above may affect how these pages render.

  8. If your destination is Save as PDF, click Save. If your destination is a printer, click Print.

Explore additional analytics and optimization options.

  • Topics: Analyze conversation trends by topic.
  • Conversations: View individual conversations from report metrics.
  • CSAT: Configure customer satisfaction surveys.
  • Knowledge: Manage articles used by your AI Agent.
  • Actions: Create and monitor API integrations.
  • Coaching: Improve AI Agent responses based on report insights.