name: blind-review description: Blind peer review for code - multiple agents critique independently, then synthesize privately to avoid groupthink triggers:
- /blind-review
- blind review
- peer review
- multi-agent review
Blind Peer Review Skill
Based on: Harvard+Stanford+CMU "LLM Review" paper - blind peer review improves quality while isolation protects originality.
Key insight: When agents openly see each other's work, they copy and converge. Blind review + private synthesis prevents groupthink.
FULLY AUTONOMOUS EXECUTION
Claude executes ALL phases without user intervention. Never ask user to run commands.
Phase 1: Identify Review Target
Determine what to review:
- If user specifies files/PR → use those
- If recent changes exist → review uncommitted changes
- If neither → ask user what to review
# Get files to review
git diff --name-only HEAD~1 # Recent commit
git diff --name-only # Uncommitted changes
Phase 2: Parallel Independent Reviews (BLIND)
Launch 3 reviewer agents in parallel using a SINGLE message with multiple Task tool calls.
CRITICAL: Each agent works in ISOLATION - they do NOT see each other's prompts or outputs.
Reviewer 1: Architecture Strategist
Prompt: Review this code from an architectural perspective. Focus on:
- Component boundaries and responsibilities
- Coupling and cohesion
- Design pattern usage
- Scalability concerns
Do NOT suggest implementation details. Only identify structural issues.
Files to review: [FILE_LIST]
Reviewer 2: Pattern Recognition Specialist
Prompt: Analyze this code for patterns and anti-patterns. Focus on:
- Code duplication
- Naming consistency
- Common anti-patterns (god objects, shotgun surgery, etc.)
- Adherence to codebase conventions
Do NOT suggest architecture changes. Only identify pattern issues.
Files to review: [FILE_LIST]
Reviewer 3: Domain Expert (project-specific)
Prompt: Review this code for domain-specific correctness. For Random Timer:
- Timer logic accuracy
- Redux state management patterns
- React Native/Expo best practices
- Theme system usage
Do NOT suggest pattern changes. Only identify domain issues.
Files to review: [FILE_LIST]
Execution
// SINGLE message with 3 parallel Task calls
Task({
subagent_type: "compound-engineering:review:architecture-strategist",
prompt: "[Architecture review prompt with files]",
description: "Blind architecture review"
})
Task({
subagent_type: "compound-engineering:review:pattern-recognition-specialist",
prompt: "[Pattern review prompt with files]",
description: "Blind pattern review"
})
Task({
subagent_type: "general-purpose",
prompt: "[Domain review prompt with files]",
description: "Blind domain review"
})
Phase 3: Collect Critiques (NO CROSS-POLLINATION)
Store each agent's output separately. Do NOT share outputs between phases.
critique_1 = [Architecture Strategist output]
critique_2 = [Pattern Recognition output]
critique_3 = [Domain Expert output]
Phase 4: Private Synthesis
A SINGLE synthesis agent receives all critiques and produces final recommendations.
Key: The synthesis agent sees critiques but reviewers never saw each other's work.
Prompt: You are synthesizing 3 independent code reviews into actionable improvements.
CRITIQUE 1 (Architecture):
{critique_1}
CRITIQUE 2 (Patterns):
{critique_2}
CRITIQUE 3 (Domain):
{critique_3}
Your task:
1. Identify overlapping concerns (high priority)
2. Identify unique insights from each reviewer
3. Resolve any contradictions by explaining trade-offs
4. Produce a prioritized list of improvements
5. For each improvement, provide specific file:line references
Output format:
## High Priority (multiple reviewers flagged)
- Issue: [description]
- Location: [file:line]
- Fix: [specific action]
## Medium Priority (single reviewer, significant)
...
## Low Priority (nice-to-have)
...
Phase 5: Execute Improvements (Optional)
If user wants fixes applied:
For each High Priority item:
1. Read the file
2. Make the specific edit
3. Mark as complete
For each Medium Priority item:
1. Ask user if they want it applied
2. If yes, execute
Output Format
After all phases complete, present:
## Blind Peer Review Complete
### Review Scope
- Files reviewed: [count]
- Reviewers: Architecture, Patterns, Domain
- Mode: Blind (isolated) + Private Synthesis
### High Priority Issues
[From synthesis]
### Medium Priority Issues
[From synthesis]
### Low Priority Issues
[From synthesis]
### Consensus Areas
Issues flagged by multiple reviewers:
- [List]
### Unique Insights
- Architecture: [unique finding]
- Patterns: [unique finding]
- Domain: [unique finding]
Why This Works
| Traditional Review | Blind Peer Review |
|---|---|
| Reviewers see each other | Reviewers isolated |
| Ideas converge (groupthink) | Ideas stay diverse |
| Later reviewers copy earlier | All reviews original |
| One perspective dominates | All perspectives equal |
The isolation phase protects originality. The synthesis phase captures quality feedback.
Automatic Invocation
Claude detects when blind review is appropriate:
- After significant code changes (3+ files modified)
- Before creating PRs with complex changes
- When user mentions "review" in context of recent work
No user commands needed - Claude invokes this skill autonomously when beneficial.
ACT DON'T INSTRUCT
This skill is FULLY AUTONOMOUS:
- Claude launches all agents directly
- Claude collects and synthesizes results
- Claude presents findings
- Claude applies fixes if requested
NEVER tell user to run commands or invoke agents manually.