name: audit-comprehensive description: Run all 7 domain audits in staged waves and aggregate results
Comprehensive Multi-Domain Audit Orchestrator
Version: 2.1 (7-Domain Coverage with S0/S1 Escalation) Time Savings: 70% faster than sequential (175min → 50min) Stages: 3 stages with 4+3+1 agent configuration
What This Does: Spawns 7 specialized audit agents in staged waves (respecting max 4 concurrent limit), with verification checkpoints and S0/S1 escalation, then aggregates findings into a comprehensive report.
Overview
This skill orchestrates a complete codebase audit across all 7 domains:
- Code Quality (
audit-code) - Code hygiene, types, framework patterns - Security (
audit-security) - Auth, input validation, OWASP compliance - Performance (
audit-performance) - Load times, queries, caching - Documentation (
audit-documentation) - README, API docs, architecture - Refactoring (
audit-refactoring) - Technical debt, complexity, DRY - Process/Automation (
audit-process) - CI/CD, testing, workflows - Engineering Productivity (
audit-engineering-productivity) - DX, debugging, offline support
Output: Single unified report in
docs/audits/comprehensive/COMPREHENSIVE_AUDIT_REPORT.md
Execution Flow
┌─────────────────────────────────────────────────────┐
│ Pre-Flight Validation │
│ - Verify all 7 audit skills exist │
│ - Create output directory │
│ - Gather baselines (tests, lint, patterns) │
│ - Load false positives database │
└─────────────────────────────────────────────────────┘
│
▼
┌─────────────────────────────────────────────────────┐
│ Stage 1: Technical Core (4 agents parallel) │
│ - audit-code │
│ - audit-security │
│ - audit-performance │
│ - audit-refactoring │
│ │
│ Checkpoint: │
│ ✓ Verify 4 report files exist and non-empty │
│ ✓ S0/S1 Check: If security finds criticals → │
│ notify user before proceeding │
└─────────────────────────────────────────────────────┘
│
▼
┌─────────────────────────────────────────────────────┐
│ Stage 2: Supporting (3 agents parallel) │
│ - audit-documentation │
│ - audit-process │
│ - audit-engineering-productivity │
│ │
│ Checkpoint: │
│ ✓ Verify 3 report files exist and non-empty │
└─────────────────────────────────────────────────────┘
│
▼
┌─────────────────────────────────────────────────────┐
│ Stage 3: Aggregation (sequential) │
│ - audit-aggregator │
│ │
│ Checkpoint: │
│ ✓ Verify COMPREHENSIVE_AUDIT_REPORT.md exists │
└─────────────────────────────────────────────────────┘
│
▼
┌─────────────────────────────────────────────────────┐
│ Post-Audit │
│ - Update AUDIT_TRACKER.md │
│ - Display final summary │
│ - Recommend next steps │
└─────────────────────────────────────────────────────┘
Pre-Flight Validation
Step 0: Episodic Memory Search (Session #128)
Before running audits, search for context from past audit sessions:
// Search for past comprehensive audits
mcp__plugin_episodic -
memory_episodic -
memory__search({
query: ["comprehensive audit", "findings", "patterns"],
limit: 5,
});
// Search for specific domain context
mcp__plugin_episodic -
memory_episodic -
memory__search({
query: ["security audit", "S0", "critical"],
limit: 5,
});
Why this matters:
- Compare against previous audit findings
- Identify recurring issues (may indicate architectural debt)
- Avoid flagging known false positives
- Track improvement/regression trends
Use findings to:
- Note which S0/S1 issues from past audits are still open
- Identify patterns that keep appearing (root cause needed)
- Set context for aggregator on known false positives
Step 1: Verify Skills Exist
Check that all 7 audit skills are available:
ls -1 .claude/skills/audit-*/SKILL.md | wc -l
# Should return 7
If not all present, notify user which audits are missing and ask whether to proceed with available audits only.
Step 2: Create Output Directory
mkdir -p docs/audits/comprehensive
Step 2.5: Verify Output Directory (CRITICAL)
Before running ANY agent, verify AUDIT_DIR is valid:
AUDIT_DIR="docs/audits/comprehensive"
AUDIT_PATH=$(realpath "${AUDIT_DIR}" 2>/dev/null || echo "${AUDIT_DIR}")
# Check for empty, root path, or path traversal attempts
if [ -z "${AUDIT_DIR}" ] || [ "${AUDIT_PATH}" = "/" ] || [[ "${AUDIT_DIR}" == ".."* ]]; then
echo "FATAL: Invalid or unsafe AUDIT_DIR"
exit 1
fi
echo "Output directory: ${AUDIT_DIR}"
Why: Context compaction can cause variable loss. Always verify before agent launches.
Step 3: Run Baseline Checks
Gather current metrics to provide to all audits:
# Test count
npm test 2>&1 | grep -E "Tests:|passing|failed" | head -5
# Lint status
npm run lint 2>&1 | tail -10
# Pattern compliance
npm run patterns:check 2>&1 | head -20
Store results in docs/audits/comprehensive/baseline.txt for reference.
Step 4: Load False Positives
Read docs/audits/FALSE_POSITIVES.jsonl to pass to aggregator (prevents
duplicate flagging of known false positives).
Stage 1: Technical Core Audits (4 Parallel)
Launch 4 agents IN PARALLEL using Task tool with run_in_background: true:
| Agent | Skill | Output File |
|---|---|---|
| 1A | audit-code | audit-code-report.md |
| 1B | audit-security | audit-security-report.md |
| 1C | audit-performance | audit-performance-report.md |
| 1D | audit-refactoring | audit-refactoring-report.md |
Why these 4 first:
- Core technical analysis
- Security findings needed for S0/S1 escalation check
- Respects max 4 concurrent agents (CLAUDE.md Section 6.3)
Display Initial Status:
🚀 Comprehensive Audit Started
━━━━━━━━━━━━━━━━━━━━━━━━━━━━
Stage 1: Technical Core (4 parallel)
⏳ Code Quality
⏳ Security
⏳ Performance
⏳ Refactoring
Stage 2: Supporting (waiting)
⏸️ Documentation
⏸️ Process/Automation
⏸️ Engineering Productivity
Stage 3: Aggregation (waiting)
⏸️ Aggregator
Estimated time: 45-50 minutes
(vs 175 minutes if run sequentially - 70% faster!)
Stage 1 Checkpoint (MANDATORY)
Before proceeding to Stage 2, perform these checks:
1. Verify output files exist:
for f in audit-code-report.md audit-security-report.md audit-performance-report.md audit-refactoring-report.md; do
if [ ! -s "docs/audits/comprehensive/$f" ]; then
echo "❌ MISSING: $f - re-run agent"
else
echo "✅ $f exists"
fi
done
2. S0/S1 Security Escalation Check:
grep -cE "\bS0\b|\bS1\b" docs/audits/comprehensive/audit-security-report.md
If S0/S1 findings exist, display:
⚠️ SECURITY ESCALATION
━━━━━━━━━━━━━━━━━━━━━━━
Security audit found critical/high findings.
These should be reviewed before continuing.
S0 Critical: X findings
S1 High: Y findings
Options:
1. Review security findings now (recommended for S0)
2. Continue with remaining audits
3. Stop and address security issues first
What would you like to do?
3. Display Stage 1 Summary:
✅ Stage 1 Complete (Technical Core)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━
✅ Code Quality (X findings)
✅ Security (X findings, Y critical)
✅ Performance (X findings)
✅ Refactoring (X findings)
Proceeding to Stage 2...
Stage 2: Supporting Audits (3 Parallel)
Launch 3 agents IN PARALLEL using Task tool with run_in_background: true:
| Agent | Skill | Output File |
|---|---|---|
| 2A | audit-documentation | audit-documentation-report.md |
| 2B | audit-process | audit-process-report.md |
| 2C | audit-engineering-productivity | audit-engineering-productivity-report.md |
Why these in Stage 2:
- Supporting audits that can use Stage 1 context
- Lower priority than technical core
- Completes the full 7-domain coverage
Stage 2 Checkpoint (MANDATORY)
1. Verify output files exist:
for f in audit-documentation-report.md audit-process-report.md audit-engineering-productivity-report.md; do
if [ ! -s "docs/audits/comprehensive/$f" ]; then
echo "❌ MISSING: $f - re-run agent"
else
echo "✅ $f exists"
fi
done
2. Display Stage 2 Summary:
✅ Stage 2 Complete (Supporting)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━
✅ Documentation (X findings)
✅ Process/Auto (X findings)
✅ Engineering Productivity (X findings)
All 7 audits complete. Proceeding to aggregation...
Stage 3: Aggregation Phase
Launch Aggregator Agent
Use Task tool to spawn audit-aggregator agent:
Task({
subagent_type: "audit-aggregator",
description: "Aggregate and deduplicate audit results",
prompt: `
Read all 7 audit reports from docs/audits/comprehensive/
Perform:
1. Deduplicate findings (same file:line across multiple audits → merge)
2. Identify cross-cutting patterns (files appearing in 3+ audits)
3. Priority ranking (severity × cross-domain count × effort)
4. Generate executive summary with top 20 findings
Output to: docs/audits/comprehensive/COMPREHENSIVE_AUDIT_REPORT.md
`,
});
Expected Output:
docs/audits/comprehensive/COMPREHENSIVE_AUDIT_REPORT.md(unified report)
Wait for aggregator to complete (typically 3-5 minutes)
Stage 3 Checkpoint (MANDATORY)
After aggregator completes, verify:
if [ ! -s "docs/audits/comprehensive/COMPREHENSIVE_AUDIT_REPORT.md" ]; then
echo "❌ Aggregation failed - report not generated"
echo "Individual reports still available for manual review"
else
echo "✅ Comprehensive report generated"
fi
Final Report Structure
The COMPREHENSIVE_AUDIT_REPORT.md should contain:
Executive Summary
- Total unique findings (after deduplication)
- Severity breakdown (S0: X, S1: Y, S2: Z, S3: W)
- Top 3 cross-domain insights
- Recommended fix order
- Effort estimate (total hours)
Priority-Ranked Findings (Top 20)
Table format:
| Rank | ID | Severity | Domains | File:Line | Description | Effort |
|---|---|---|---|---|---|---|
| 1 | COMP-001 | S0 | 3 | auth.ts:45 | Missing auth check | E1 |
| ... | ... | ... | ... | ... | ... | ... |
Cross-Domain Insights
Examples:
- "Files X, Y, Z appear in 4+ audits → Comprehensive refactor needed"
- "Security + Performance overlap: 12 findings where fixing one helps both"
- "Documentation gaps align with code complexity hotspots"
Full Findings (Deduplicated)
Complete table of all findings grouped by severity, with links to original audit reports.
Appendix
- Links to individual audit reports
- Baseline metrics snapshot
- False positives excluded (count)
Post-Audit (MANDATORY)
After aggregation completes, you MUST update tracking:
1. Update AUDIT_TRACKER.md
Add an entry to each of the 6 category tables in docs/AUDIT_TRACKER.md:
| Date | Session | Commits Covered | Files Covered | Findings | Reset Threshold |
|---|---|---|---|---|---|
| {TODAY} | Comprehensive | Full codebase | All | Session #{N} - [report link] | ✅ (all) |
2. Update Threshold Summary Table
In the "Current Thresholds" section, update all 6 categories:
- Set "Last Audit" to today's date with "(Comprehensive)"
- Reset "Commits Since" to 0
- Reset "Files Since" to 0
3. Update Multi-AI Thresholds
In the "Multi-AI Audit Thresholds" section:
- Update "Total commits" reset date to today
- Update "Time elapsed" to "0 days (comprehensive audit {TODAY})"
This step ensures npm run review:check correctly shows no triggers after the
audit.
Triage & Roadmap Integration (MANDATORY)
After TDMS intake completes, triage new items into the roadmap:
1. Review New Items
Check the newly added DEBT-XXXX items:
# View recent additions (last 50 items by ID)
tail -50 docs/technical-debt/MASTER_DEBT.jsonl | jq -r '[.id, .severity, .category, .title[:60]] | @tsv'
2. Priority Scoring
Beyond S0-S3 severity, consider these factors for prioritization:
| Factor | Weight | Description |
|---|---|---|
| Severity | 40% | S0=100, S1=50, S2=20, S3=5 |
| Cross-domain | 20% | Items flagged by multiple audits get +50% |
| Effort inverse | 20% | E0=4x, E1=2x, E2=1x, E3=0.5x (quick wins first) |
| Dependency | 10% | Blockers for other items get +25% |
| File hotspot | 10% | Files with 3+ findings get +25% |
Priority Score Formula:
score = (severity × 0.4) × (cross_domain_mult × 0.2) × (effort_inv × 0.2) × (dep_mult × 0.1) × (hotspot_mult × 0.1)
3. Track Assignment
New items are auto-assigned based on category + file patterns:
| Category | File Pattern | Track |
|---|---|---|
| security | * | Track-S |
| performance | * | Track-P |
| process | * | Track-D |
| refactoring | * | M2.3-REF |
| documentation | * | M1.5 |
| code-quality | scripts/, .claude/ | Track-E |
| code-quality | .github/ | Track-D |
| code-quality | tests/ | Track-T |
| code-quality | functions/ | M2.2 |
| code-quality | components/, lib/, app/ | M2.1 |
View current assignments:
cat docs/technical-debt/views/unplaced-items.md
4. Update ROADMAP.md
For S0/S1 items that need immediate attention:
## Track-S: Security Technical Debt
- [ ] DEBT-0875: Firebase credentials written to disk (S1) **NEW**
- [ ] DEBT-0876: Missing App Check validation (S1) **NEW**
For bulk items by track:
- [ ] DEBT-0869 through DEBT-0880: Process automation gaps (S2, bulk)
5. Consistency Check
Verify all references are valid:
node scripts/debt/sync-roadmap-refs.js --check-only
Reports:
- Orphaned refs (in ROADMAP but not in MASTER_DEBT)
- Unplaced items (in MASTER_DEBT but not in ROADMAP)
- Status mismatches (marked done but not RESOLVED)
6. Review Cadence
| Trigger | Action |
|---|---|
| After comprehensive audit | Full triage of all new items |
| After single-domain audit | Triage items in that category only |
| Weekly (if no audits) | Check unplaced-items.md for drift |
| Before sprint planning | Review S0/S1 items for inclusion |
Completion
Display Final Summary:
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━
🎯 COMPREHENSIVE AUDIT COMPLETE
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━
📊 Results Summary:
• 142 raw findings → 97 unique (45 merged)
• S0 Critical: 3
• S1 High: 24
• S2 Medium: 42
• S3 Low: 28
🔍 Cross-Domain Insights:
• 8 files need comprehensive refactor (4+ audits)
• 12 security/performance overlaps
• 5 documentation gaps in complex areas
📄 Full Report:
docs/audits/comprehensive/COMPREHENSIVE_AUDIT_REPORT.md
⏱️ Total Time: ~45 minutes
(vs 150 minutes sequential - saved ~105 minutes!)
🎯 Recommended Next Steps:
1. Review top 20 priority findings
2. Create GitHub issues for S0/S1
3. Plan refactor for hotspot files
Error Handling
If Individual Audit Fails:
- Continue with remaining audits in the same stage
- Mark failed audit in status display (❌)
- Note failure in final report
- Suggest running failed audit individually for debugging
If Aggregator Fails:
- All individual reports still available
- User can manually review 6 separate reports
- Suggest creating GitHub issue for aggregator failure
If All Audits Fail:
- Check baseline environment (tests passing, lint working)
- Check for system issues (disk space, memory)
- Suggest running single audit first to isolate issue
Context Recovery
If context compacts mid-audit, resume from last completed checkpoint:
Determine Current State
echo "=== Checking audit progress ==="
ls -la docs/audits/comprehensive/*.md 2>/dev/null | wc -l
Recovery Matrix
| Files Found | State | Resume Action |
|---|---|---|
| 0-3 reports | Stage 1 incomplete | Re-run missing Stage 1 audits |
| 4 reports | Stage 1 complete | Start Stage 2 |
| 5 reports | Stage 2 incomplete | Re-run missing Stage 2 audit |
| 6 reports, no COMPREHENSIVE | Stage 2 complete | Run Stage 3 (aggregator) |
| COMPREHENSIVE exists | Complete | Run post-audit only |
Resume Commands
Stage 1 incomplete: Re-run only missing audits:
# Check which are missing
for audit in code security performance refactoring; do
[ ! -f "docs/audits/comprehensive/audit-${audit}-report.md" ] && echo "Missing: $audit"
done
Stage 2 incomplete: Run documentation and/or process audits as needed.
Stage 3: Run aggregator on existing 6 reports.
Usage Examples
Quarterly Health Check:
/audit-comprehensive
Pre-Release Audit:
/audit-comprehensive
After Major Refactor:
/audit-comprehensive
Focused Audit (Not Comprehensive):
Use individual skills instead:
/audit-security (25 min - when you only need security review)
/audit-code (30 min - when you only need code quality)
Notes
- Staged Execution: 3 stages (4+2+1 agents) respects CLAUDE.md max 4 concurrent limit
- Time Estimate: ~45 minutes (vs 150min sequential = 70% savings)
- S0/S1 Escalation: Security findings checked after Stage 1 before proceeding
- Checkpoints: Each stage verifies outputs exist and are non-empty
- Context Recovery: Can resume from any checkpoint after context compaction
- Output Consistency: All audits use same severity (S0-S3) and effort (E0-E3) scales
Future Enhancements
- Incremental Audits: Only re-run audits for changed domains
- Custom Audit Subset:
--audits code,securityto run subset - Confidence Scoring: Weight findings by audit confidence levels
- Trend Analysis: Compare against previous comprehensive audits
- Auto-Issue Creation: Create GitHub issues for S0/S1 findings automatically
Related Skills
/audit-code- Individual code quality audit/audit-security- Individual security audit/audit-performance- Individual performance audit/audit-documentation- Individual documentation audit/audit-refactoring- Individual refactoring audit/audit-process- Individual process/automation audit/audit-aggregator- Standalone aggregation (if you have existing reports)
Documentation References
Before running this audit, review:
TDMS Integration (Required)
- PROCEDURE.md - Full TDMS workflow
- MASTER_DEBT.jsonl - Canonical debt store
- All individual audits automatically run TDMS intake after completion
Documentation Standards (Required)
- JSONL_SCHEMA_STANDARD.md - Output format requirements and TDMS field mapping (used by aggregator)
- DOCUMENTATION_STANDARDS.md - 5-tier doc hierarchy
Version History
| Version | Date | Description |
|---|---|---|
| 2.1 | 2026-02-03 | Added Triage & Roadmap Integration section with priority scoring formula |
| 2.0 | 2026-02-02 | Staged execution (4+2+1), S0/S1 escalation, checkpoints, context recovery |
| 1.0 | 2026-01-28 | Initial version - flat parallel execution of all 6 audits |