Model Optimizer 🧠
AI model management and cost optimization for agent systems.
Triggers
Use this skill when:
- "optimize model costs"
- "track AI spending"
- "best model for task"
- "model performance"
- "switch models"
- "ai provider comparison"
- "token usage tracking"
- "model health check"
- Agent needs to select optimal models based on cost/performance
- Managing multiple AI provider APIs
What It Does
Cost Tracking
- Monitor token usage and costs across providers (OpenAI, Anthropic, Google, local)
- Daily/weekly/monthly spend reports
- Cost per task type analysis
- Budget alerts and limits
Performance Monitoring
- Response time tracking
- Success/failure rates
- Quality metrics by task type
- Provider availability monitoring
Smart Model Selection
- Route tasks to optimal models based on cost/performance profiles
- Fallback chains when providers are down
- Task-specific model recommendations
- Load balancing across providers
Usage Analytics
- Token efficiency analysis
- Provider cost comparison
- Peak usage patterns
- ROI by model/task
Usage
Track Model Usage
# Log a model call
node scripts/log-usage.js --provider openai --model gpt-4 --input-tokens 150 --output-tokens 75 --cost 0.002 --task "code-review"
# Get cost report
node scripts/cost-report.js --period weekly --provider all
# Check current spend vs budget
node scripts/budget-check.js
Model Selection
# Get best model for task
node scripts/recommend-model.js --task "code-generation" --budget-limit 0.01 --max-latency 3000
# Check provider health
node scripts/health-check.js --provider anthropic
# Get fallback chain
node scripts/fallback-chain.js --primary openai --task summarization
Analytics
# Usage analytics
node scripts/analytics.js --metric efficiency --period 7d
# Compare providers
node scripts/compare-providers.js --metric cost-per-token --timeframe 30d
# Export usage data
node scripts/export-usage.js --format csv --start 2024-01-01
Configuration
Create config/models.json:
{
"providers": {
"openai": {
"models": {
"gpt-4": {
"input_cost": 0.03,
"output_cost": 0.06,
"context_limit": 8192,
"strengths": ["reasoning", "code"],
"latency_avg": 2000
},
"gpt-3.5-turbo": {
"input_cost": 0.0015,
"output_cost": 0.002,
"context_limit": 4096,
"strengths": ["speed", "cost"],
"latency_avg": 800
}
}
},
"anthropic": {
"models": {
"claude-3-opus": {
"input_cost": 0.015,
"output_cost": 0.075,
"context_limit": 200000,
"strengths": ["analysis", "long-context"],
"latency_avg": 3000
}
}
},
"local": {
"models": {
"llama-70b": {
"input_cost": 0,
"output_cost": 0,
"context_limit": 4096,
"strengths": ["privacy", "cost"],
"latency_avg": 5000
}
}
}
},
"budgets": {
"daily": 10.0,
"weekly": 50.0,
"monthly": 200.0
},
"task_profiles": {
"code-generation": {
"priority": "quality",
"max_cost": 0.05,
"preferred_providers": ["openai", "anthropic"]
},
"summarization": {
"priority": "speed",
"max_cost": 0.01,
"preferred_providers": ["openai", "local"]
},
"analysis": {
"priority": "quality",
"max_cost": 0.10,
"preferred_providers": ["anthropic", "openai"]
}
}
}
Files Created
data/usage.jsonl- Usage logsdata/performance.jsonl- Performance metricsdata/costs.jsonl- Cost trackingconfig/models.json- Model configurations
Integration
OpenClaw Hook
// In your agent code
const optimizer = require('./skills/model-optimizer/scripts/optimizer');
async function callModel(task, prompt) {
const recommendation = await optimizer.recommend(task);
const result = await callAPI(recommendation.provider, recommendation.model, prompt);
await optimizer.logUsage(recommendation, result);
return result;
}
Budget Monitoring
# Add to cron
0 */6 * * * node /path/to/model-optimizer/scripts/budget-check.js --alert-webhook $WEBHOOK_URL
Example Output
$ node scripts/cost-report.js --period weekly
Model Cost Report (Last 7 Days)
================================
Total Spend: $12.45
By Provider:
- OpenAI: $8.20 (66%)
- Anthropic: $4.25 (34%)
- Local: $0.00 (0%)
Top Models:
1. gpt-4: $6.50 (458 calls)
2. claude-3-opus: $4.25 (127 calls)
3. gpt-3.5-turbo: $1.70 (1,245 calls)
By Task Type:
- code-generation: $7.80 (63%)
- analysis: $3.20 (26%)
- summarization: $1.45 (11%)
Efficiency Metrics:
- Avg cost per call: $0.027
- Token efficiency: 92%
- Failed calls: 2.3%
Budget Status: $37.55 remaining this week
Dependencies
- Node.js 16+
- No external packages required (pure Node.js)
Author
Axiom 🔬
@AxiomBot · github.com/0xAxiom