Skip to main content
The Conversation Monitoring evaluator enables you to monitor conversations and identify issues in them that may impact the quality of customer interactions. Our monitoring system can flag a conversation as having potential quality issues as determined by our quality evaluators.

How it works

Our monitoring system uses quality evaluators to assess each turn for adherence to configured tasks and knowledge. When evaluators detect issues, they:
  • Flag the conversation
  • Highlight problematic utterances within the conversation
  • Provide rationale for flagging.
Once a conversation is flagged, you can review it in the Conversations interface by applying the appropriate filters.

Identifying quality issues

When quality issues are detected, the conversation review interface provides the following features:
  • Inline indicators: Flagged messages appear with visual indicators directly in the conversation flow. This allows the reviewer to quickly identify potential issues.
  • Quality Tab: A dedicated tab in the “Conversations” interface provides detailed information about each detected utterance and acts as a centralized location for quality-related insights. This includes:
    • List of all the detected messages in the conversation
    • Specific turn(s) that were flagged
    • Reason provided by the conversation monitoring system for flagging the message
  • Customizable flaging: You can change the severity level of the flagged messages (e.g., from “major” to “critical”) or dismiss them if they are false positives. This helps refine the monitoring system over time.
Quality tab
The features above help reviewers quickly identify and understand quality issues in conversations, enabling them to take appropriate actions to address them.

Next steps

After identifying quality issues in conversations, you can take the following next steps to improve the overall performance of your GenerativeAgent:
  • Audit AI-driven response quality
  • Identify regressions from new tasks, prompts, or configurations
  • Generate insights for evaluator training, task design, and knowledge updates
  • Improve automation accuracy and reduce escalations from model errors
Consider exploring the following evaluators for more in-depth analysis: