Skip to main content
Monitor workflow tool and model performance — track run success rates, average response times, node executions, latency, request volume, scaling, and credit usage. Use workflow tool-specific and model-specific filters for trend analysis, and drill into individual traces for run-level inputs, outputs, and metadata.

Tool Analytics Dashboard

The Tool Analytics Dashboard tracks performance of deployed tools and their versions. Use it to monitor run success and failure rates, average response times (ART), and event-based node executions. Navigation: Go to SettingsMonitoringAnalyticsTool Analytics.

Key Metrics

The dashboard displays account-level and workflow tool-specific metrics.

Account-Level Metrics

These metrics summarize activity across all workflow tools in your account for the selected period.
MetricDescription
Number of toolsTotal deployed tools, including all versions managed by all account users.
Number of runsTotal successful and failed runs across all deployed tools.
Number of times nodes are executedEvent-driven execution counts for all node types. For node-level details, see the Nodes Execution widget.

Tool-Specific Metrics

These metrics appear in the performance widgets for the selected tool and version.
  • Tool Runs: Success and failure trends for tool runs.
  • Avg. Tool Response Time: ART to requests for the selected tool version.
  • Nodes Execution: Event-based execution graph for all node types. Executions are counted up to your account’s rate limit threshold.

Filter Dashboard Data

Global Timeline Filters

Select the time period for all dashboard metrics. The default is 7 days. Custom date selection supports single dates and consecutive date ranges up to one year from the current date.

Tool Performance Filters

These filters apply only to tool-specific metrics and work alongside a global timeline selection.
  • Tool Name: If two tools share the same name, they appear with their deployment timestamps.
  • Tool Version: Use this to compare testing versus production performance. By default, all versions are shown.

Performance Widgets

The dashboard has three widgets for tool-specific metrics.

Tool Runs

This widget shows successful and failed tool runs as two line graphs for the selected period.
  • Y-axis: Total Number of Requests (auto-scaled).
  • X-axis: Date.
  • Successful runs appear in green; failed runs in red.
Hover over a data point to view total, successful, and failed request counts at that date and time. tooltip feature
  • The widget shows data only for periods the tool was deployed. For example, if a tool was active for 4 days and undeployed for 3 days in a 7-day view, only the 4 active days appear.
  • The graph shows curves only when requests are processed. No data points appear for hours or days with no activity.
Use this widget to:
  • Monitor success versus failure rates and identify failure patterns across the same or different dates.
  • Identify peak and low performance periods across multiple date ranges.
  • Detect sudden spikes in failures or performance drops with the hourly view.
  • Compare performance across tool versions.

Avg. Tool Response Time

This widget shows a line graph of the ART for each request the selected tool version handles during the selected period.
  • Y-axis: Avg. Response Time (auto-scaled).
  • X-axis: Date.
  • A blue line represents ART values.
Single day Hover over a data point to see the date and time, ART for that hour, and number of requests processed. The widget also shows Overall ART — the aggregated ART across all hours in the selected day. overall art Date range Hover over a data point to see the date, ART for that day, and number of requests processed. The widget also shows Overall ART — the aggregated ART across all days in the selected date range. overall art for date range
The system calculates the ART for all requests within each hour (hourly ART), combines these for each day (daily ART), then aggregates daily ART values across the selected date range to compute the overall ART.
Use this widget to:
  • Identify bottlenecks by spotting sudden ART spikes that indicate delays or inefficiencies.
  • Track optimization progress by monitoring ART trends before and after tool changes.
  • Compare ART across tool versions to determine the most responsive option.

Nodes Execution

This widget shows a line graph of nodes executed per event run for the selected tool version and period. For information about node types, see Nodes.
  • Y-axis: Number of Nodes Executed, fixed to the maximum rate limit for your account.
  • X-axis: Date.
  • A blue line represents node executions.
  • A red line marks the maximum rate limit threshold.
Rate limit The widget displays data up to your account’s subscribed rate limit. For example, if your subscription allows 8,000 node executions, the graph shows a maximum of 8,000 nodes. When the limit is reached, an error message appears. rate limit Use this widget to:
  • Identify overused nodes or inefficient tools by visualizing execution trends.
  • Stay within subscription thresholds by tracking node executions against your rate limit.
  • Detect bottlenecks and resolve tool issues using node execution patterns.

Model Analytics Dashboard

The Model Analytics Dashboard tracks performance for fine-tuned, open-source, and external models in your account. It measures:
  • Model latency and response times.
  • Request success and failure rates.
  • Model scaling patterns and usage.
  • Credit consumption for deployments and fine-tuning.
Navigation: Go to SettingsMonitoringAnalyticsModel Analytics.

Key Metrics

The dashboard displays account-level and model-specific metrics.

Account-Level Metrics

These metrics summarize credit usage and model counts across your account for the selected period.
MetricDescription
Credits consumed in deploymentCredits deducted from your account for deploying models.
Credits consumed in fine-tuningCredits deducted from your account for fine-tuning models.
Number of Deployed ModelsTotal models deployed in your account.
Number of Fine-tuned ModelsTotal models fine-tuned in your account.

Model-Specific Metrics

Available metrics vary by model type.
MetricFine-tuned and Open-sourceExternal
Model LatencyPer-request latency for a 24-hour or 7-day period. For longer ranges, average daily latency is shown. See Model Latency.Same as fine-tuned and open-source.
RequestsSuccess and failure trends, total requests, and average credits per request. See Requests.Same, but without average credits per request.
Model Scaling and UsageReplica count and hardware configuration during deployment. See Model Scaling and Usage.Not available.
TokensNot available.Input and output token bar chart for the selected period. See Tokens.

Filter Dashboard Data

Global Timeline Filters

Select the time period for all dashboard metrics. The default is 7 days. Custom date selection supports single dates and consecutive date ranges up to one year from the current date.

Model Performance Filters

These filters apply only to model-specific metrics. Available filters depend on the selected model type. If multiple deployments share the same model name, they appear with their deployment timestamps.
  • Fine-tuned and Open-source models: Select the model name, deployment name, and deployment version. By default, all versions are shown.
  • External models: Select the model name and third-party service connection name.

Performance Widgets

To load data in the widgets, select a date or date range, model type tab, and model filters to load data.
  • Widgets show data only for periods the model was deployed. For example, if a model was active for 4 days and undeployed for 3 days in a 7-day view, only the 4 active days appear.
  • The graph shows curves only when requests are processed. No data points appear for hours or days with no activity.

Model Latency

Available for Fine-tuned, Open-source, and External models. This widget shows a line graph of the selected model version’s latency during the selected period.
  • Y-axis: Latency in milliseconds (auto-scaled).
  • X-axis: Date.
  • A blue line shows per-request latency for 24-hour or 7-day periods. For longer date ranges, the graph shows average daily latency.
Hover over a data point to view the latency for a specific request (24-hour or 7-day periods) or the average latency for a specific date and time (longer ranges). hover over info Use this widget to:
  • Identify peak usage and low-efficiency periods by tracking latency over time.
  • Detect sudden latency spikes that indicate performance issues.
  • Compare deployed model versions to identify the best-performing option.
  • Correlate latency trends with credit usage for resource optimization.

Requests

Available for Fine-tuned, Open-source, and External models. This widget shows successful and failed requests as two line graphs for the selected period.
  • Y-axis: Number of Requests (auto-scaled).
  • X-axis: Date.
  • Successful requests appear in green; failed requests in red.
  • Total Requests: Combined count of successful and failed requests for the selected period.
  • Avg. credits per request: Total credits consumed divided by total responses generated.
Avg. credits per request is not shown for external models.
Hover over a data point to view the count of successful or failed requests at that date and time, and the total count for the full selected period. requests hover info Use this widget to:
  • Monitor success versus failure rates and identify failure patterns.
  • Detect sudden spikes in failures using the hourly view.
  • Compare request performance across model versions.

Model Scaling and Usage

Available for Fine-tuned and Open-source models only. This widget shows a step graph of the number of replicas with a specific hardware configuration deployed for the selected model version.
  • Y-axis: Number of Replicas (adjusts based on deployed replicas).
  • X-axis: Date.
  • Each upward step represents additional replicas being generated; each downward step represents replicas being undeployed.
Multiple Hardware is shown when replicas with more than one hardware configuration are deployed across a date range.
Hover over a data point to view the timestamp, deployment version name, replica count, and hardware configuration at that point. model scaling hover info Maximum replicas limit For a single-day selection, a red line indicates the maximum number of replicas configured during model deployment. max replicas Use this widget to:
  • Track replica deployments over time to identify usage spikes and optimize hardware consumption.
  • Spot inefficiencies in replica usage and reduce unnecessary deployment costs.
  • Monitor scaling against your account’s replica limits to stay within subscription thresholds.

Tokens

Available for External models only. This widget shows a stacked bar graph of input and output tokens for the selected period.
  • Y-axis: Number of Tokens (in thousands, K; auto-scaled).
  • X-axis: Date.
  • Input and output tokens appear in two shades in a stacked bar, showing them as parts of the total tokens used.
  • The widget shows the total sum of input and output tokens.
input output tokens Hover over a data point to view the timestamp and input and output token counts at that point. tokens hover info Use this widget to:
  • Assess model efficiency by comparing input versus output token counts.
  • Track token-based credit consumption and optimize usage.
  • Identify processing bottlenecks and usage patterns over time.