Operational Dashboards

The five operational dashboards provide detailed, actionable insights for day-to-day management.

Last updated 11 days ago

1. Tools Dashboard

Path: /tools

Purpose: Analyze tool usage, performance, and reliability.

KPI Cards

Tool Calls

Total calls during the period

Count of all tool invocations

Success Rate

Percentage of calls completed successfully

Successful calls ÷ Total calls × 100

Average Duration

Average response time in seconds

Sum of durations ÷ Total calls

Total Cost

Cost of tool calls (USD)

Sum of tool call costs

Daily Tool Usage Chart

A stacked bar chart showing three datasets per day:

  • Total calls — All tool invocations

  • Successful calls — Calls that completed without errors

  • Errors — Failed calls

Use this to:

  • Track daily tool workload

  • Identify days with high error rates

  • Correlate tool usage with business events

Top Tools

A ranked table of the most used tools (by instance), showing:

Tool Name

Name of the tool instance

Calls

Number of invocations

Success Rate

Percentage of successful calls

Avg Duration

Average response time (seconds)

Use this to:

  • Identify your most critical tools

  • Spot tools with low success rates that need debugging

  • Optimize slow tools for better user experience

Usage by Agent

Shows which Mates use which tools.

Displays:

  • Mate name

  • Tools used by that Mate

  • Number of calls per tool

Use this to:

  • Understand tool dependencies

  • Identify Mates that rely heavily on external tools

  • Plan for tool maintenance or deprecation

Top Users

Shows which users trigger the most tool calls.

Displays:

  • User name

  • Number of tool calls triggered

  • Top tools used

Use this to:

  • Identify power users of specific tools

  • Provide targeted training or support

  • Understand usage patterns

Tool Performance

Two distribution charts:

By Status

A donut chart showing:

  • Success — Completed successfully

  • Processed — Completed with warnings

  • Error — Failed

  • Running — Currently executing (rare in historical data)

By Duration

A bar chart categorizing calls by response time:

Instant

< 0.5s

Fast

0.5–2s

Medium

2–5s

Slow

5–10s

Very slow

10s

Use this to:

  • Set performance SLAs

  • Identify slow tools that need optimization

  • Understand user experience impact

Usage Intensity

Concentration metrics:

Tools per message

Average per agent message

Tools per conversation

Average per session

Tools per user

Average per person

Distinct tools

Usage diversity

Use this to:

  • Understand how tool-dependent your Mates are

  • Identify opportunities to reduce tool calls for cost savings

  • Benchmark against best practices

Insights Panel

Automated alerts based on current data:

"High error rate"

10% of calls failing

Investigate failing tools immediately

"Response time to optimize"

20% of calls >5s

Review slow tools for optimization

"Excellent reliability"

<5% error rate

System is healthy — maintain current practices

2. Attachments Dashboard

Path: /attachments

Purpose: Analyze file usage, processing performance, and storage.

KPI Cards

Files Uploaded

Total files over the period

Count of all file uploads

Total Storage

Storage consumed (GB or MB)

Sum of file sizes

Tokens Extracted

Tokens parsed from files

Sum of tokens extracted via RAG

Error Rate

Files with processing errors

Files with errors ÷ Total files × 100

Files Uploaded per Day Chart

A dual-axis bar chart showing:

  • Left axis (blue bars) — Number of files uploaded per day

  • Right axis (green line) — Volume in MB per day

Use this to:

  • Track file upload trends

  • Identify peak upload days

  • Plan storage capacity

File Types

A donut chart showing distribution by category:

  • PDF — PDF documents

  • Images — JPG, PNG, GIF, etc.

  • Text — TXT, MD, etc.

  • Documents — DOCX, ODT, etc.

  • Spreadsheets & CSV — XLSX, CSV, etc.

  • Presentations — PPTX, etc.

  • Code & Data — JSON, YAML, etc.

  • Archives — ZIP, TAR, etc.

  • Videos — MP4, etc.

  • Audio — MP3, WAV, etc.

Use this to:

  • Understand what types of content users are working with

  • Optimize processing pipelines for common file types

  • Identify unsupported formats that users are trying to upload

Size Distribution

Files grouped by size:

Tiny

< 10 KB

Small

10–100 KB

Medium

100 KB – 1 MB

Large

1–10 MB

Very large

10 MB

Displayed as: A bar chart showing the count of files in each category.

Use this to:

  • Identify storage optimization opportunities

  • Set file size limits

  • Plan for large file processing capacity

Active Conversations

A ranked list of conversations with the most file attachments.

Displays:

  • Conversation ID or title

  • Number of files attached

  • Total storage used

Use this to:

  • Identify file-heavy use cases

  • Provide targeted support for users working with many files

  • Understand how files are used in conversations

Content Extraction

Processing performance metrics:

Tokens per file

Average tokens extracted per file

Pages per file

Average page count (for PDFs)

Parsing speed

KB processed per second

Total pages processed

Sum of all pages extracted

Use this to:

  • Benchmark processing performance

  • Identify bottlenecks in file processing

  • Optimize RAG pipelines

Usage Intensity

Adoption metrics:

Files per conversation

Average files per session

KB per file

Average file size

Conversations with files

Count of sessions with ≥1 file

Active workspaces

Workspaces with file uploads

File types used

Number of distinct file types

Use this to:

  • Understand file usage patterns

  • Identify opportunities to promote file-based workflows

  • Measure the diversity of file usage

3. Errors Dashboard

Path: /errors

Purpose: Monitor errors, identify root causes, and assess impact.

KPI Cards

Message Errors

Messages with errors during the period

Count of messages with error status

Tool Errors

Failed tool calls

Count of tool calls with error status

Global Error Rate

Percentage of operations with errors

(Message errors + Tool errors) ÷ Total operations × 100

Impacted Users

Users who encountered at least one error

Count of unique users with ≥1 error

Error Evolution Chart

An area chart showing the trend over time for:

  • Message errors — LLM errors (red area)

  • Tool errors — Tool call failures (orange area)

Use this to:

  • Track error trends

  • Identify error spikes

  • Measure the impact of fixes or changes

Error Sources

A donut chart splitting errors between:

  • Messages — LLM errors (model failures, rate limits, etc.)

  • Tools — Tool call failures (API errors, timeouts, etc.)

Shows: The error rate for each source.

Use this to:

  • Understand the primary source of errors

  • Prioritize debugging efforts

  • Allocate resources to the most problematic area

Error Rate by Workspace

Identifies which workspaces have the highest error rates.

Displays:

  • Workspace name

  • Total errors

  • Error rate (%)

Use this to:

  • Identify problematic workspaces

  • Provide targeted support

  • Investigate workspace-specific issues (e.g., misconfigured Mates)

Error Rate by Tool

Shows reliability metrics per tool.

Displays:

  • Tool name

  • Total calls

  • Errors

  • Error rate (%)

Use this to:

  • Identify unreliable tools

  • Debug tool configurations

  • Communicate with tool providers about issues

Errors by Agent

LLM and tool errors per Mate — helps identify problematic AI configurations.

Displays:

  • Mate name

  • Message errors

  • Tool errors

  • Total error rate (%)

Use this to:

  • Identify Mates with high error rates

  • Review Mate instructions or model configurations

  • Disable or fix problematic Mates

Errors by Model

Shows which LLM models generate the most errors.

Displays:

  • Model name

  • Total errors

  • Error rate (%)

Use this to:

  • Identify unreliable models

  • Switch to more stable models

  • Report issues to model providers

Most Impacted Users

A table showing for each user:

User

Name and avatar

Errors

Number of errors encountered

Mates Affected

Number of Mates with errors

Tools Affected

Number of tools with errors

Total Interactions

Total messages sent

Messages Sent

Messages sent by this user

Default view: Top 10 impacted users Expandable: Click "View all users" to see the complete list

Use this to:

  • Identify users who need support

  • Understand the user experience impact

  • Prioritize fixes based on user impact

Error Impact

Summary metrics:

Impacted conversations

Conversations with ≥1 error

Concerned agents

Mates with ≥1 error

Tools in error

Tools with ≥1 failure

Average error rate

Overall error rate (%)

Use this to:

  • Quantify the scope of errors

  • Report to stakeholders

  • Set error reduction goals

4. Credits Dashboard

Path: /credits

Purpose: Track Polar quota consumption and native connection costs.

KPI Cards

Credits Consumed

Polar units consumed (current state)

Sum of consumed credits across all meters

Credits Remaining

Polar units available

Sum of remaining credits across all meters

LLM Credit Cost

LLM token cost via native connections

Cost of tokens consumed through allmates.ai-managed connections

Tool Credit Cost

Tool call cost via native connections

Cost of tool calls through allmates.ai-managed connections

Note: Credits are tracked via Polar meters. Each meter tracks a specific type of usage (e.g., LLM tokens, tool calls).

Polar Meters

Visual progress bars showing credit quota status per meter:

For each meter:

  • Meter name (e.g., "LLM Tokens", "Tool Calls")

  • Global usage bar — Visual representation of consumed vs. allocated credits

  • Consumed — Credits used

  • Allocated — Total credits available

  • Remaining — Credits left

Status alerts:

  • "Near limit" — >80% consumed (orange warning)

  • "Over limit" — >100% consumed (red alert)

Use this to:

  • Monitor quota consumption

  • Avoid service interruptions

  • Plan for quota increases

Credit Cost Evolution Chart

A dual-axis line chart showing:

  • Left axis (blue line) — LLM cost over time

  • Right axis (green line) — Tool cost over time

Use this to:

  • Track spending trends

  • Identify cost spikes

  • Correlate costs with business events

Cost by LLM Model

Breakdown of credit costs by model.

Displays:

  • Model name

  • Credits consumed

  • Percentage of total LLM cost

Use this to:

  • Understand which models are most expensive

  • Optimize model selection

  • Negotiate pricing with providers

Credit Cost by Tool

Native tool cost breakdown with call counts.

Displays:

  • Tool name

  • Credits consumed

  • Number of calls

  • Cost per call

Use this to:

  • Identify expensive tools

  • Optimize tool usage

  • Budget for tool costs

5. User Activity Dashboard

Path: /my-activity

Purpose: Detailed overview of an individual user's activity.

How to Use

  1. Select a user from the dropdown at the top of the page

  2. Use the search field to find a user by name or username

  3. The page loads all metrics for the selected user and period

KPI Cards

Messages Sent

Messages sent by this user

Conversations

Conversations this user participated in

Mates Used

Distinct Mates this user interacted with

Tokens Generated

Tokens consumed by this user's activity

Tool Calls

Tool calls triggered by this user

Active Days

Days with at least one interaction

Estimated Cost

Cost estimate for this user's consumption

Errors

Errors encountered by this user

Daily Activity Chart

A dual-axis chart showing:

  • Left axis (blue bars) — Messages per day

  • Right axis (green line) — Tokens per day

Use this to:

  • Track individual user engagement over time

  • Identify usage patterns

  • Spot anomalies or drops in activity

Favorite Mates

The user's most solicited Mates with request counts.

Displays:

  • Mate name

  • Number of requests

Use this to:

  • Understand which Mates this user finds most valuable

  • Provide personalized training or support

  • Identify opportunities to introduce new Mates

Top Tools

The tools most triggered by this user.

Displays:

  • Tool name

  • Number of calls

Use this to:

  • Understand this user's workflow

  • Optimize tool configurations for this user

  • Identify training needs

Tokens by Model

LLM consumption distribution across models.

Displays:

  • Model name

  • Tokens consumed

  • Percentage of total

Use this to:

  • Understand which models this user prefers

  • Optimize model selection for this user's use cases

  • Budget for this user's costs

Workspaces

Activity breakdown by workspace.

Displays:

  • Workspace name

  • Messages sent

  • Conversations

  • Tokens consumed

Use this to:

  • Understand where this user is most active

  • Identify collaboration patterns

  • Provide workspace-specific support