Lesson: Monitoring, Metrics & Dashboards · Topic 2 of 2
Lesson summary
Now that production calls are flowing into Bluejay, you need three things: evaluation criteria that match your business goals, dashboards that show you trends at a glance, and alerts that wake you up when quality drops. This topic covers all three — Custom Metrics, Dashboards, and Alerts — and walks you through setting up each one.Objectives
- Create Custom Metrics with clear, specific evaluation criteria
- Use Metrics Lab to prototype and refine metrics before promoting them
- Build a dashboard to visualize agent quality over time
- Set up threshold-based alerts with Slack notifications
Video walkthrough
Video coming soon. Follow the written walkthrough below in the meantime.
Walkthrough
Create Custom Metrics
Custom Metrics are the evaluation rules Bluejay uses to score every conversation — both simulation runs and production calls. They answer specific questions about the conversation, like “Did the agent confirm the appointment date?” or “Was the disclaimer read aloud?”To create one, go to Custom Metrics in the sidebar and click New Metric. You will need to fill in:
For the full reference, see Key Concepts → Custom Metrics.
- Name: A short label (e.g., “Greeting check”)
- Prompt: The question Bluejay should answer about the conversation
- Response type: How the answer should be formatted
| Response type | What it returns | Example use case |
|---|---|---|
| Pass / Fail | Binary yes/no result | ”Did the agent read the required disclaimer?” |
| Yes / No | Binary with different labeling | ”Did the caller express frustration?” |
| Qualitative | Rating scale (poor / fair / good / excellent) | “Overall conversation quality” |
| Quantitative | Numeric value | ”Number of times the agent interrupted” |
| JSON | Structured data | ”Extract the order details mentioned” |
| Enum | One value from a predefined list | ”Call outcome: resolved / escalated / dropped” |
Write effective metrics
The quality of your metrics determines the quality of your scores. Follow these three rules:
| Rule | Bad example | Good example |
|---|---|---|
| Be specific | ”Was the call good?" | "Did the agent confirm the appointment date and time?” |
| Be observable | ”Was the agent helpful?" | "Was the greeting delivered within the first 10 seconds?” |
| Be actionable | ”Rate the call" | "Did the agent offer an escalation path when unable to help?” |
Prototype with Metrics Lab
Metrics Lab lets you test a metric against real conversations before you promote it to production scoring. This way you can refine the prompt and see how it scores without affecting live data.The workflow:
- Draft a metric in Metrics Lab
- Run it against a set of past conversations (side-by-side comparison)
- Adjust the wording until scores match your expectations
- Promote the metric — it will now score all new conversations automatically
Build a dashboard
Dashboards give you a bird’s-eye view of agent quality. They combine data from both simulations and production calls into one place.What you will see on a dashboard:
- Health scores — Aggregate pass rates for your Custom Metrics
- Trend lines — Sparklines showing how metrics change over time
- Alert badges — Visual indicators when a metric has breached a threshold
- Quick actions — Run simulations, open call logs, or drill into details
Set up alerts
Alerts notify you when a metric drops below a threshold you set. For example, you might want to know the moment your greeting-check metric falls below 80%.To create an alert:
- Go to Monitor → Alarms and click New Alarm
- Select the Custom Metric to watch
- Set the threshold (e.g., “fire when pass rate drops below 80%”)
- Choose the notification channel — Slack or email
Connect Slack for notifications
To receive alert notifications in Slack:
- Go to Settings → Integrations and find the Slack section
- Click Connect and authorize Bluejay in your Slack workspace
- When creating or editing an alarm, select the Slack channel where notifications should go
Activity
Knowledge check
What is the difference between a simulation metric and an observability metric?
What is the difference between a simulation metric and an observability metric?
There is no structural difference — Custom Metrics are the same entity. They can be used to evaluate both simulation runs and production calls. You create them once and they work everywhere.
How quickly do alerts fire?
How quickly do alerts fire?
Alerts evaluate after each call is scored. If the configured threshold is breached, the alert fires immediately and delivers to the configured channel (Slack or email).
What is Metrics Lab for?
What is Metrics Lab for?
Metrics Lab lets you draft and test a metric against real conversations before promoting it to production. This way you can refine the wording and check that scores match your expectations without affecting live scoring.