description: > Calculate a productivity score using actual Agent Monitor metrics — session completion rates, cache efficiency (cache_read vs input), compaction pressure (baseline tokens), turn velocity (turn_count / total_turn_duration_ms), tool success ratio (PreToolUse vs PostToolUse), and the workflow intelligence API's complexity and effectiveness scores.
Productivity Score
Calculate a productivity scorecard from the Agent Monitor's real data.
Input
The user provides: $ARGUMENTS
Options: "today", "this week", "last 30 days", a session ID, or "compare" for period comparison.
Data Sources
| Endpoint | Returns |
|---|---|
GET /api/analytics | Token totals (total_input, total_output, total_cache_read, total_cache_write — baselines pre-summed), tool_usage top 20, daily_events/sessions, event_types, sessions_by_status, agents_by_status, avg_events_per_session, total_subagents |
GET /api/sessions?limit=100 | Sessions with metadata JSON: thinking_blocks, turn_count, total_turn_duration_ms, usage_extras (service_tier, speed, inference_geo) |
GET /api/pricing/cost | Total cost with per-model breakdown |
GET /api/workflows/{sessionId} | 11 workflow datasets: stats, orchestration, toolFlow, effectiveness, patterns, modelDelegation, errorPropagation, concurrency, complexity, compaction, cooccurrence |
Score Components (each 0–100)
1. Completion Rate (20% weight)
From sessions_by_status:
completed / (completed + error + abandoned) × 100- Bonus for high completed-to-active ratio
- Penalty for abandoned sessions (wasted work)
2. Token Efficiency (20% weight)
From analytics tokens (baselines are pre-summed into totals):
- Cache hit rate:
total_cache_read / (total_cache_read + total_input) × 100- Above 60% = excellent, below 30% = poor
- Output concentration:
total_output / total_input— 0.3–0.8 is balanced
3. Tool Effectiveness (20% weight)
From event_types:
- Success ratio: Count
PostToolUse/ CountPreToolUse— should be ~1.0; gap = tool failures - API error rate: Count
APIError/ total events — should be near 0 - From workflow
effectivenessdata: subagent completion rates, task success per type
4. Velocity (20% weight)
From session metadata:
- Turns per session: average
turn_countacross sessions - Turn speed: average
total_turn_duration_ms / turn_count— lower = faster - Events per session: from
avg_events_per_sessionin analytics overview - Thinking depth: average
thinking_blocks— more thinking = more thorough (neutral metric)
5. Cost Efficiency (20% weight)
From pricing:
- Cost per completed session:
total_cost / completed_sessions - Cost trend: comparing current period to previous (decreasing = improving)
- Model optimization: sessions using expensive models (Opus) for tasks subagents handle with Haiku/Sonnet
Overall Score
Weighted sum → letter grade:
- A+ (95-100), A (90-94), B+ (85-89), B (80-84), C+ (75-79), C (70-74), D (60-69), F (<60)
Output Format
═══════════════════════════════════════
PRODUCTIVITY SCORE: 87/100 (B+)
═══════════════════════════════════════
Completion Rate ████████░░ 80/100
Token Efficiency █████████░ 92/100
Tool Effectiveness████████░░ 85/100
Velocity █████████░ 88/100
Cost Efficiency █████████░ 90/100
═══════════════════════════════════════
Then: top 3 strengths, top 3 improvement areas with actionable steps, and period comparison if available.