name: agent-audit
description: >
Audit your AI agent setup for performance, cost, and ROI. Scans OpenClaw config, cron jobs,
session history, and model usage to find waste and recommend optimizations.
Works with any model provider (Anthropic, OpenAI, Google, xAI, etc.).
Use when: (1) user says "audit my agents", "optimize my costs", "am I overspending on AI",
"check my model usage", "agent audit", "cost optimization", (2) user wants to know which
cron jobs are expensive vs cheap, (3) user wants model-task fit recommendations,
(4) user wants ROI analysis of their agent setup, (5) user says "where am I wasting tokens".
Scan your entire OpenClaw setup and get actionable cost/performance recommendations.
1. Scans config — reads OpenClaw config to map models to agents/tasks
2. Analyzes cron history — checks every cron job's model, token usage, runtime, success rate
3. Classifies tasks — determines complexity level of each task
4. Calculates costs — per agent, per cron, per task type using provider pricing
5. Recommends changes — with confidence levels and risk warnings
6. Generates report — markdown report with specific savings estimates
python3 {baseDir}/scripts/audit.py
Options:
python3 {baseDir}/scripts/audit.py --format markdown # Full report (default)
python3 {baseDir}/scripts/audit.py --format summary # Quick summary only
python3 {baseDir}/scripts/audit.py --dry-run # Show what would be analyzed
python3 {baseDir}/scripts/audit.py --output /path/to/report.md # Save to file
Classify each task into complexity tiers:
| Tier | Examples | Recommended Models |
|------|----------|-------------------|
| Simple | Health checks, status reports, reminders, notifications | Cheapest tier (Haiku, GPT-4o-mini, Flash, Grok-mini) |
| Medium | Content drafts, research, summarization, data analysis | Mid tier (Sonnet, GPT-4o, Pro, Grok) |
| Complex | Coding, architecture, security review, nuanced writing | Top tier (Opus, GPT-4.5, Ultra, Grok-2) |
Classification signals:
For each task where the model tier doesn't match complexity:
⚠️ RECOMMENDATION: Downgrade "Knox Bot Health Check" from opus to haiku
Current: anthropic/claude-opus-4 ($15/M input, $75/M output)
Suggested: anthropic/claude-haiku ($0.25/M input, $1.25/M output)
Reason: Simple status check averaging 300 output tokens
Estimated savings: $X.XX/month
Risk: LOW — task is simple pattern matching
Confidence: HIGH
Output a clean markdown report with:
1. Overview — total agents, crons, monthly spend estimate
2. Per-agent breakdown — model, usage, cost
3. Per-cron breakdown — model, frequency, avg tokens, cost
4. Recommendations — sorted by savings potential
5. Total potential savings — monthly estimate
6. One-liner config changes — exact model strings to swap
See [references/model-pricing.md](references/model-pricing.md) for current pricing across all providers.
Update this file when prices change.
See [references/task-classification.md](references/task-classification.md) for detailed heuristics
on how tasks are classified into complexity tiers.