Docs
GitHub Open Dashboard
v0.1.0 Updated Feb 2026

HiveBoard — Insights Guide #

Version: 0.2.0 Last updated: 2026-02-15

38 questions, four narrative moments — what your fleet data tells you that you should know.


Table of Contents #

  1. Overview
  2. Toolbar and Controls
  3. Moment 1: The Glance
  4. Moment 2: The Investigation
  5. Moment 3: The Optimization
  6. Moment 4: The Review
  7. Question Reference
  8. Data Sources

1. Overview #

The Insights page (insights.html) is an intelligence briefing for your AI agent fleet. Rather than showing raw data or configurable dashboards, it organizes 38 computed questions into four narrative "moments" — each designed for a specific mindset.

The page is designed to be read top-to-bottom like a report: start with a glance at fleet health, investigate specific agents, find optimization opportunities, and end with a holistic review.

Each moment is a collapsible section. Each question within a moment is rendered as a card or visualization with a question ID (Q1–Q38) shown in the corner for reference.

Key distinction from Analytics: Analytics lets you explore data freely across six analytical dimensions. Insights tells you what the data means — it's opinionated, narrative-driven, and includes automated detectors and health scoring.


2. Toolbar and Controls #

Element Description
Range selector Choose the analysis window: 1 hour, 6 hours, 24 hours (default), 7 days
Agent filter Optional filter to scope insights to a specific agent
Refresh button Manual refresh with spinning animation. Reloads all data from APIs
Last updated Timestamp of the most recent data fetch

Unlike Analytics (which auto-refreshes every 60s), Insights refreshes on demand via the Refresh button. This is intentional — Insights is designed for deliberate review, not continuous monitoring.


3. Moment 1: The Glance #

Section title: "The Glance — Fleet Status at a Glance"

This moment answers: "Is everything OK? Do I need to act?" It's designed to be scannable in under 5 seconds.

Q1: Fleet Status Bar #

A horizontal bar segmented by agent status (processing, idle, stuck, error, waiting). Each segment shows the count and is proportionally sized. Below the bar, a legend labels each status with its count.

Read it as: Green/blue segments = healthy. Any red segments = action needed.

Q2: Fleet Vitals #

Four big-number cards:

Card Shows Good sign
Total Agents Count of registered agents Matches expectations
Active Now Agents currently processing Non-zero during work hours
Success Rate Fleet-wide task success percentage ≥ 90%
Tasks Completed Total completed tasks in the range Consistent with workload

Q3: Cost Snapshot #

Three cards for cost awareness:

Card Shows
Total Spend Fleet-wide LLM cost for the range
Cost per Task Average LLM cost per task
Projected Monthly Current spend extrapolated to 30 days

Q4: Throughput Sparkline #

A mini bar chart showing task completion volume over time. Helps spot drops or spikes in activity.

Q5: Latest Activity Feed #

The 10 most recent events across the fleet, showing timestamp, agent name, event type, and summary. Includes a live badge if events are very recent.


4. Moment 2: The Investigation #

Section title: "The Investigation — Agent & Task Deep Dive"

This moment answers: "Let me dig into a specific agent — what's its story?"

Agent Selector #

A prominent selector bar at the top of this section lets you choose which agent to investigate. Styled with the accent color to distinguish it from the global agent filter.

Q6: Agent Status Card #

A detailed card for the selected agent showing:

  • Status badge (Processing/Idle/Stuck/Error)
  • Heartbeat age with colored indicator
  • Current task ID (if processing)
  • Last event type and time

Q7: Agent Performance Summary #

Key performance numbers for the selected agent in the current range:

  • Tasks completed / failed
  • Success rate
  • Average duration
  • Total LLM cost
  • LLM calls per task average

Q8: Agent vs Fleet Comparison #

Side-by-side comparison of the selected agent against fleet averages:

Metric Agent Fleet Avg
Success rate
Avg duration
Cost per task
Throughput

Values better than fleet average are highlighted green; worse values are highlighted red.

Q9: Agent Task Breakdown #

Distribution of the selected agent's tasks by type, shown as horizontal bars with counts and percentages.

Q10–Q16: Extended Agent Analysis #

Additional cards covering:

  • Q10: Error breakdown for this agent (by type)
  • Q11: LLM model usage distribution
  • Q12: Top actions/tools by invocation count
  • Q13: Queue depth and pipeline status
  • Q14: Cost trend over the selected range
  • Q15: Duration trend over the selected range
  • Q16: Task completion timeline

5. Moment 3: The Optimization #

Section title: "The Optimization — Cost & Reliability"

This moment answers: "Where can I improve? What should I fix or optimize?"

Q17: Cost Leaderboard #

Ranking of agents by total cost, with the most expensive highlighted. Shows cost, call count, and share of fleet total.

Q18: Cost per Task Ranking #

Agents ranked by average cost per task. Useful for finding agents that are expensive per unit of work (even if their total spend is modest due to low volume).

Q19: Model Cost Comparison #

Breakdown of cost by LLM model across the entire fleet. Helps identify whether switching models could save money.

Q20: Token Efficiency #

Analysis of input/output token ratios. Agents with very high input-to-output ratios may be sending oversized prompts.

Q21: Slowest Tasks #

A table of the slowest tasks in the fleet, showing task ID, agent, duration, and status. Helps identify performance bottlenecks.

Q22: Error Hotspots #

Agents ranked by error count with error rate percentages. Includes breakdown by error type.

Q23: Retry Analysis #

If agents use retries, shows retry frequency and which tasks/actions trigger the most retries.

Q24: Action Performance Ranking #

Actions ranked by success rate and average duration. Low success rates and high durations are highlighted as optimization opportunities.

Q25: Idle Time Analysis #

How much time agents spend idle vs. working. High idle percentages may indicate over-provisioning.

Q26: Queue Health #

Analysis of queue depths and item ages across agents. Long queues or stale items indicate capacity problems.

Smart Detectors (Q27–Q28) #

Automated checks that flag specific operational problems:

Detector What it checks Status
Stuck Agent Detector Any agents with stale heartbeats OK / Warning / Error
Error Spike Detector Sudden increase in error rate OK / Warning
Cost Anomaly Detector Cost significantly above normal OK / Warning
Queue Backlog Detector Growing queue depths OK / Warning
Idle Fleet Detector All agents idle when they shouldn't be OK / Warning
Success Rate Detector Fleet success rate below threshold OK / Warning

Each detector shows as a card with a colored left border: green (OK), amber (Warning), red (Error). The card includes a description of what was checked and the current finding.


6. Moment 4: The Review #

Section title: "The Review — Trends & Fleet Health"

This moment answers: "Stepping back — how is the fleet doing overall, and is it ready to scale?"

Q29–Q35: Metric Summaries #

Quick-reference cards for fleet-wide totals:

Card Shows
Q29: Total tasks Completed + failed count
Q30: Total LLM calls Fleet-wide LLM invocation count
Q31: Total cost Fleet-wide LLM spend
Q32: Total tokens Input + output token totals
Q33: Average duration Fleet-wide mean task duration
Q34: Fleet error count Total errors in the range
Q35: Agent uptime Summary of agent availability

Q36: Trend Summary #

Computes directional trends (up/down/flat) for four key metrics over the selected time range:

Metric Good direction
Throughput Up ↑
Errors Down ↓
Cost Down ↓
Duration Down ↓

Each trend shows a percentage change and a colored indicator. A verdict banner at the bottom summarizes:

Verdict Condition Color
"Trends look healthy. Keep it up." ≥ 3 of 4 trends favorable Green
"Mixed signals — some trends need attention." 2 of 4 favorable Amber
"Multiple concerning trends. Review details above." ≤ 1 favorable Red

Requires at least 4 timeseries data points to compute. Shows "Not enough data points" otherwise.

Q37: Health Gauge #

A visual gauge (0-100) representing overall fleet health. The score starts at 100 and subtracts for problems:

Deduction Points
Each stuck agent -15 per agent
Each agent in error state -10 per agent
Success rate below 90% -(90 - rate) × 0.5
Active issues -5 per issue (max -20)
No agents registered -50

The gauge displays as a semi-circular SVG with color coding:

Score Label Color
80–100 Healthy Green
50–79 Needs Attention Amber
0–49 Critical Red

Below the gauge, a deduction breakdown shows exactly what reduced the score (e.g., "1 stuck (-15) · Success rate 82% (-4)"). If no deductions: "No deductions — everything looks great!"

Q38: Scale Readiness Checklist #

A pass/fail checklist evaluating whether the fleet is ready to scale:

Check Pass condition
All agents online Every agent's heartbeat is within threshold
No stuck agents Zero stuck agents
Success rate ≥ 90% Fleet-wide success rate meets threshold
No active issues Zero unresolved issues
Cost/task < $1.00 Average cost per task is under a dollar
Multi-agent fleet At least 2 agents registered
Recent activity At least one event in the current range

Summary banner:

Result Condition Color
"Ready to scale" All checks pass Green
"Almost ready" All but 1–2 checks pass Amber
"Not ready yet" 3+ checks fail Red

7. Question Reference #

Quick lookup of all 38 questions by ID:

ID Moment What it answers
Q1 Glance Fleet status distribution
Q2 Glance Fleet vitals (agents, active, success rate, tasks)
Q3 Glance Cost snapshot (total, per-task, projected monthly)
Q4 Glance Throughput sparkline
Q5 Glance Latest activity feed
Q6 Investigation Selected agent status
Q7 Investigation Selected agent performance
Q8 Investigation Agent vs fleet comparison
Q9 Investigation Agent task type breakdown
Q10-16 Investigation Extended agent analysis
Q17 Optimization Cost leaderboard
Q18 Optimization Cost per task ranking
Q19 Optimization Model cost comparison
Q20 Optimization Token efficiency
Q21 Optimization Slowest tasks
Q22 Optimization Error hotspots
Q23 Optimization Retry analysis
Q24 Optimization Action performance ranking
Q25 Optimization Idle time analysis
Q26 Optimization Queue health
Q27-28 Optimization Smart detectors
Q29-35 Review Metric summaries
Q36 Review Trend summary with verdict
Q37 Review Health gauge (0-100 score)
Q38 Review Scale readiness checklist

8. Data Sources #

The Insights page pulls from seven API endpoints:

Endpoint Used by
GET /v1/agents Q1, Q2, Q6, Q37, Q38, Smart Detectors
GET /v1/metrics Q2, Q3, Q4, Q7, Q8, Q29-Q35, Q36, Q37, Q38
GET /v1/cost Q3, Q17, Q18, Q19, Q20
GET /v1/events Q5, Q38
GET /v1/pipeline Q13, Q26, Q37, Q38, Smart Detectors
GET /v1/tasks Q9, Q21
GET /v1/insights/* Q10-Q12, Q14-Q16, Q22-Q24

All data is fetched on page load and when the Refresh button is clicked. The range selector controls the time window for all time-scoped queries.