Skip to content

Understanding Tool Monitor

The Tool Monitor capabilities are shown on the following tabs:

  • All runs: It shows comprehensive data on all tool run instances and provides a comprehensive record of all the endpoint calls made to the tool.

  • Model runs: This tab shows specific data on AI node run instances, focusing on endpoint calls made to AI nodes within the tool.

Both tabs include summary metrics at the top of the page, including total runs/requests, average response times (P90 and P99), and failure rates.

All runs and Model runs tabs

All Runs

The All runs tab provides the following information for each tool run:

  • Run ID: The unique identifier for the flow.
  • Status: The current state of the request. It displays one of the following statuses:
    • In Progress: The request is being processed.
    • Waiting: The request is outside of Agent Platform and is awaiting a response from the connected system (typically for API nodes in Async mode).
    • Success: The request has been completed successfully.
    • Failed: The request was not completed successfully.
  • Response time: The duration the tool takes to complete a request and provide an output.
  • Nodes executed: The total number of nodes executed in the run.
  • Start time: The time when the request is initiated.
  • End time: The time the response is received.
  • Source: Based on whether the tool was run using an Agentic App or an API endpoint, the following information is displayed:
    • Agentic App - The name of the Agentic App used to run the tool.
    • API Key: The API key name used to execute the tool. (This is the name provided in the API keys page when you created a new API key. If you did not provide a name, the system uses the default name ‘Secret Key’. You can have multiple API keys for a single tool).

Note

When there is a nested (multi-level) tool calling, the immediate tool (parent) is displayed as the source.

You can also see the following metrics on the top of the page:

  • TOTAL RUNS: The total number of tool runs.
  • RESPONSE TIME: The average response time of all the tool runs. It is measured by the following two metrics:
    • P90: This metric represents the response time below which 90% of the requests fall.
    • P99: This metric represents the response time below which 99% of the requests fall.
  • FAILURE RATE: The percentage of failed runs.

These metrics are dependent on the selected date range, filters, and search criteria. When you apply a search, the metrics on the top of the page will be updated to reflect the specific request. Essentially, the displayed top metrics will vary based on the filters you apply.

Model Runs

Each AI node in the tool is recorded as a separate request in the Model runs tab of the Tool monitor page. For example, if three AI nodes are used in a tool, three separate requests for each of the three AI nodes are displayed.

If your tool currently does not have any AI nodes, this section will remain empty. Once you add AI nodes, the Model runs will begin to reflect here.

The Model runs tab provides the following information for each AI node call:

  • Request ID: The unique identifier of the AI node request.
  • Status: The current state of the request. It displays one of the following statuses:
    • In Progress: The request is being processed.
    • Waiting: The request is outside of Agent Platform and is awaiting a response from the connected system (typically for API nodes in Async mode).
    • Success: The request has been completed successfully.
    • Failed: The request was not completed successfully.
  • Node name: The name of the AI node.
  • Model name: The model that is used for the AI node.
  • Response time: The amount of time taken by the AI node to complete the request.
  • Start time: The time when the AI node has started its execution.
  • End time: The time when the AI node has completed its execution.

You can also see the following metrics on the top of the page:

  • TOTAL REQUESTS: The total number of AI node requests.
  • RESPONSE TIME: The average response time of all the AI node requests. It is measured by the following two metrics:
    • P90: This metric represents the response time below which 90% of the requests fall.
    • P99: This metric represents the response time below which 99% of the requests fall.
  • FAILURE RATE: The percentage of instances in which the AI node has failed in execution.

These metrics are dependent on the selected date range, filters, and search criteria. When you apply a search, the metrics on the top of the page will be updated to reflect the specific request. Essentially, the displayed top metrics will vary based on the filters you apply.

Viewing Detailed Run Information

Clicking each row in either the All runs tab or the Model runs tab opens a detailed view panel on the right. This view is similar to the Run dialog panel on the Tool flow canvas. Learn more.

Viewing detailed run information

The panel displays the following details:

  • Run ID/Request ID: The unique identifier for the flow.
  • Response Time: The duration the tool takes to complete a request and provide an output.
  • Debug icon: Clicking this icon displays the debug log details.
  • Input: The Input section displays the input sent to the tool.
  • Flow log: The flow log section displays the information of each node.

    • Success: Displays the log as in the debug panel.
    • Failure: Displays failure details as in the debug panel.

      For AI nodes, when you expand the node you can see the information related to each node along with the scanner information.

  • Output section: The Output section displays the tool's output (for successful runs). You can copy the output and view tokens.

    Viewing detailed run information

Understanding the Impact of Timeouts on Tool Endpoints

The impact of timeouts on tool endpoints depends on whether the process is synchronous (Sync) or asynchronous (Async). Sync requests are handled and fulfilled immediately, while Async requests may pause and show a ‘Waiting’ status until a response is received. If the response time is longer or the timeout is set to infinite, the system will wait indefinitely until the external system responds.

Below are the four scenarios showing how timeouts affect the tool endpoint, along with the corresponding status on the Tool monitoring page:

Tool 'Sync' & API node 'Sync':

  • Request immediately fulfilled, no specific message to the endpoint.
  • 'In-progress' status while running.

Tool 'Sync' & API node 'Async' (API node timeout < Sync timeout):

  • Tool API retrieves data, flow executes as 'In-progress' status, and the response is sent.
  • External requests: Tool execution is paused awaiting external’s systems response with 'Waiting' status, resumes to 'In-progress' when tool execution resumes.

Tool 'Async' & API node 'Sync':

  • Tool executes, and the response is sent to the callback URL.
  • 'In-progress' status while flow is running.

Tool 'Async' & API node 'Async' (API node timeout < Tool Async timeout OR both are set to infinite):

  • External requests: Tool execution is paused awaiting external’s systems response with 'Waiting' status, resumes to 'In-progress' when tool execution resumes.
  • If the external system tries the same callback URL again, it will be notified that the request has already been fulfilled.

The timeout settings affect how long the system waits for responses and how it handles retries, ensuring proper status updates and communication with external systems. For more information on configuring timeouts, see Configure a tool and API Node.

Searching and Filtering Information

Use the search box in the top right corner of the Tool monitor page to find specific runs or calls based on keywords.

Use the calendar option to search for runs or calls from a specific time period. You can filter your search results by time period, whether it’s something from the last day, week, month, or year.

Steps to use the time-based search:

  1. Click the calendar button in the top right corner of the Tool monitor page.

  2. Select a predefined time range or set custom dates.

  3. Click Apply to update the results.

Custom Filters

Use the filter option to filter the information displayed in the Tool monitor dashboard by applying custom filters. These filters allow you to select specific columns, apply operators such as Is Equal To or Is Not Equal To, and then specify the desired value.

You can also add multiple filters using AND/OR operators for more precise results.

Steps to use the filters:

  1. Click the Filter icon.
  2. Click + Add filter.
  3. Select options for Column, Operator, and Value.
  4. Click Apply.

Tool Run Errors

In the All runs section, any error that occurs via the endpoint during an tool run is displayed in a separate window for the specified Run ID.

To view detailed error information, click on the corresponding tool run entry in the Tool Monitor dashboard.

An error message includes the following information:

  • HTTP status code returned by the web server as a response.
  • A message describing the error.
  • Suggestions to verify and manage the error.

Error Categories

The errors are classified as follows:

  • Authorization: An error that occurs during API key authorization of a tool.
  • Data Validation: Any discrepancy detected when validating input fields and API calls during a tool run.
  • Content Filter: Breaches of guardrail threshold limits during AI node execution.
  • Internal Server Error: Technical issues encountered with the internal server.
  • Network: Technical issues encountered with the network connectivity.

Error Scenarios

The table below lists the errors that can occur in the Tool Monitoring dashboard, including the error categories and HTTP status codes:

Error Scenario Description Category HTTP Status Code
Mandatory input field A mandatory input field is missing for the tool run. Data Validation 400 Bad Request
Invalid data type for input field An incorrect data type is provided for a field input.
Empty Input Object A field input is missing a value or has an empty value.
Large Request Payload The request payload exceeds the server's size limit. 413 Payload Too Large
Any Server side issues A technical issue caused the server to fail. Internal Server 500 Internal Server Error
Network Issues:- Request timeout on the server Temporary network or Agent Platform server connection issue. Network 408 Request Timeout
Guardrail Failure The flow execution was aborted at the AI node due to a guardrail violation, as the risk score exceeded the threshold. Content Filter 403 Forbidden