📺Monitor

Overview

The Monitor feature in ZBrain offers comprehensive oversight of your AI agents and applications by automating evaluation and performance tracking. It ensures response quality, helps identify issues proactively, and maintains optimal performance across all deployed solutions.

Monitor captures inputs and outputs from your applications, continuously evaluating responses against defined metrics at scheduled intervals. This provides real-time insights into performance, tracking success and failure rates, and highlighting patterns needing attention. The results are displayed in an intuitive interface, enabling you to quickly identification and resolution of issues and ensure consistent, high-quality AI interactions.

Key capabilities

  • Automated evaluation: Assess responses using LLM-based and non-LLM-based metrics.

  • Performance tracking: Track success/failure trends.

  • Query-level monitoring: Configure evaluations at the individual query level within a session.

  • Agent and app support: Monitor both AI apps and AI agents.

  • Input flexibility: Monitor responses for .txt, PDF, image, and other file types.

Monitor interface navigation

The monitor module consists of three main sections, accessible from the left navigation panel:

  • Events: View and manage all configured monitoring events

  • Monitor logs: Review detailed execution results and metrics

  • Event settings: Configure evaluation metrics and parameters

How to configure a monitoring event for apps

Step 1: Access the monitoring configuration

To set up monitoring for an application:

Access the app session:

  • Navigate to the Apps page

  • Click on your desired application

  • Go to the query history section of the app

  • Select a specific user session from the list to monitor (shows session ID, user info, prompt count)

Review the conversation:

  • View the session details and chat history. You can configure monitoring events at the query level. For instance, in a conversation containing multiple queries, each query can have its own monitoring event

  • Examine the response options (copy, conversation logs, edit annotation reply and feedback)

Access conversation logs:

  • Click 'Conversation Log' to see the interaction details of a specific query

  • Review status, time, and token usage

  • Check the input, output, and metadata

Enable monitoring:

  • Click the ‘Monitor’ button in the overview tab

  • Click ‘Configure now’ when prompted with ‘Added for monitoring’

Step 2: Configure event settings

You will be redirected to the Events > Monitor page. In the last status column, click ‘Configure’ to open the event settings page. On the event settings screen:

  1. Review entity information

    • Entity name: The name of your application

    • Entity type: The type of entity being monitored (e.g., App)

  2. Verify monitored content

    • Monitored input: The query or prompt being evaluated

    • Monitored output: The response being assessed

  3. Set evaluation frequency

    • Click the dropdown menu under "Frequency of evaluation"

    • Select the desired interval (Hourly, Every 30 minutes, Every 6 hours, Daily, Weekly, or Monthly)

  1. Configure evaluation conditions

    • Click ‘Add metric’ in the Evaluation Conditions section

    • Select a metric type:

LLM-based metrics

These metrics assess the quality of responses generated by large language models:

  • Response relevancy: Checks how well the response answers the user's question. Higher scores mean better alignment with the query.

  • Faithfulness: Measures whether the response accurately reflects the given context, minimizing hallucinated or incorrect information.

Non LLM-based metrics These metrics rely on algorithmic comparisons to evaluate responses without involving language models:

  • Health check: Determines whether an entity is operational and capable of producing a valid response. Note: When the health check fails, no further metric evaluations are performed for that execution.

  • Exact match: Compares the response to the expected output for an exact character-by-character match.

  • F1 score: Balances precision and recall to assess how well the response captures the expected content.

  • Levenshtein similarity: Calculates how closely two strings match based on the number of edits needed to transform one into the other.

  • ROUGE-L score: Evaluates similarity by identifying the longest common sequence of words between the generated and reference responses.

LLM as a judge metrics:

These metrics simulate human-like evaluation by using an LLM to judge qualitative aspects of a response:

  • Creativity: Rates how original and imaginative the response is in addressing the prompt.

  • Helpfulness: Assesses how well the response guides or supports the user in resolving their query.

  • Clarity: Measures how easy the response is to understand and how clearly it communicates the intended message.

  • Choose the evaluation method (is less than, is greater than, or equals to)

  • Set the threshold value (0.1 to 5.0)

  • Click ‘Add’ to save the metric

  • Set the "Mark evaluation as" dropdown to fail or success

  1. Configure notifications

  • Toggle the ‘Send Notification’ option to enable alerting for this monitoring event.

  • Click ‘+ Add a Flow from the Flow Library’. This opens the Add a Flow panel.

  • In the panel, search for the desired notification flow and select it.

  • Click the Play ▶️ button to run a delivery test.

  • If the flow succeeds, a confirmation message appears: "Flow Succeeded".

  • If the flow fails, inline error messages will be displayed, along with a link to Edit Flow for troubleshooting.

Note: Users cannot update the event settings until a valid notification flow passes the delivery test. Once the flow passes the delivery test, notifications will be sent via the chosen communication channel whenever the associated event fails.

  1. Test your configuration

    • Click the ‘Test’ button

    • Enter a test message if needed

    • Review the test results

    • Click ‘Reset’ if you want to try again

  1. Save your configuration

    • Click ‘Update’ to save and activate your monitoring event

How to configure a monitoring event for agents

Step 1: Access the monitoring setup

To set up monitoring for an agent:

Access the agent dashboard:

  • Go to the Agents page.

  • Choose a deployed agent. This opens that specific agent’s dashboard.

Enable monitoring:

  • Click on the full-screen button of ‘Agent Activity.’

  • Click the 'Monitor' button of that execution.

  • When prompted, select configure ‘Now.’

Step 2: Configure event settings

When you click on configure ‘Now,’ you’ll be redirected to the ‘Monitor’ page. Click ‘Configure’ in the ‘Last Status’ column to open the ‘Event Monitoring Settings’ page.

On the Event Monitoring screen, review the following settings, similar to the app monitoring configuration:

  • Entity Name / Type (Agent)

  • Monitored Input (e.g., a prompt or document)

  • Monitored Output (response generated by agent)

Configure the following options:

  • Evaluation Frequency (Hourly, Daily, etc.)

  • Metrics: Use LLM or non-LLM-based scoring

  • Choose the evaluation method and set the similarity threshold.

  • Outcome (mark as Success or Fail)

  • Click ‘Test’ and then ‘Update’ to save.

Once activated, the monitoring event runs automatically at the specified evaluation frequency, such as every 30 minutes, daily, or weekly.

Events dashboard

The events dashboard displays all configured monitoring events in a tabular format:

  • Entity name: The agent or application being monitored

  • Entity type: Classification (App, Agent, Playground, etc.)

  • Input: The query/input being evaluated

  • Output: The response being assessed

  • Run frequency: How often does evaluation occur

  • Last run: When the last evaluation occurred

  • Last status: Current status with the ‘Configure’ option

Use the search box and dropdown filters, Entity (App/Agent/Playground) and Status (All, Success, Failed), to quickly locate specific events. Click ‘Configure’ in the last status column to modify event settings.

Monitor logs

The monitor logs interface provides detailed performance tracking:

Key components

  1. Event information header

    • Event ID: Unique identifier for the monitoring event

    • Entity name: The agent being monitored

    • Entity type: Classification (App, Playground, etc.)

    • Frequency: How often monitoring occurs

    • Metric: Performance criteria being measured

  2. Log status visualization

    • Colored bars provide a quick visual indicator of recent execution results

    • Red bars indicate failures, green indicates successful evaluations

  3. Filtering options

    • Status dropdown: Filter by All/Success/Failed/Error status

    • Log time dropdown: Filter by active/inactive

  4. Log details table

    • Log ID: Unique identifier for each log entry

    • Log time: When the evaluation occurred

    • LLM response: The query or prompt content

    • Credits: Resource utilization

    • Cost: Associated expense

    • Metrics: Success (✅), Failure (❌)

    • Status: Outcome (Success/Failed/Error with color coding)

Setting appropriate thresholds

  • Start with conservative thresholds (0.5-0.7) and adjust based on observed performance

  • Consider your use case requirements when setting thresholds:

    • Customer-facing applications may require higher thresholds

    • Internal tools might tolerate lower thresholds

  • Regularly review and adjust thresholds as your applications evolve

Troubleshooting

  • Investigate failed evaluations by reviewing the specific LLM responses

  • Check metric scores to understand why responses did not meet thresholds

  • Adjust prompts or application configuration based on monitoring insights

User management in monitor

ZBrain Monitor supports role-based user permissions. Admins can assign access to specific users who can view and configure monitoring events.

How to manage users

  • From the Monitor page, select any monitoring event.

  • Navigate to the ‘User Management’ tab.

  • A user management panel opens with the following elements:

    • Entity name: Name of the agent/app.

    • Entity type: AGENT or APP

    • Builder: Select the builder you want to invite. The options include custom or everyone.

      • When selecting the ‘Custom’ option, you can use the search field to locate builders to invite. Enter the user’s name and click the ‘Invite’ button to assign them the Builder role.

  • Once accepted, the user will see the Monitor tab upon login and will be able to manage the assigned event.

Last updated