Prompt Manager — Quick Reference
Expert-only knowledge for transforming vague prompts into optimized, actionable ones. This guide focuses on non-obvious patterns, hard-won insights, and systematic optimization techniques.
How to Use This Guide
- Start with SKILL.md — Run the 4-phase workflow for any prompt optimization
- Reference this guide — Quick lookup for patterns, frameworks, and techniques
- Load details on-demand — Deep-dive into specific
references/files as needed
Quick Reference
Anti-Patterns (NEVER Do)
Fabrication Techniques — MoE/ToT/GoT make Claude invent fake personas instead of deepening reasoning → references/credit-killing-patterns.md#fabrication
CoT on Reasoning Models — Claude 4.5+ has native extended thinking, adding "think step-by-step" degrades quality → references/credit-killing-patterns.md#cot
Framework Name Pollution — Never output "Using CO-STAR..." or label sections with methodology → references/credit-killing-patterns.md#naming
Context-Free Optimization — Prompts for Claude Code differ from ChatGPT/API calls, context determines strategy → references/credit-killing-patterns.md#context
Vague Success Criteria — "Better", "comprehensive", "clean" lack measurability, pin to objective outcomes → references/credit-killing-patterns.md#criteria
Missing Creative Constraints — Without boundaries, creative tasks produce inconsistent results → references/credit-killing-patterns.md#constraints
Lost-in-the-Middle — Models weaken attention on middle sections of long prompts, place critical info at start/end → references/credit-killing-patterns.md#context-positioning
Ambiguous Pronouns — "It/this/that" become unclear in multi-step workflows, use specific nouns → references/credit-killing-patterns.md#pronouns
Frameworks
When to Use Which:
- CO-STAR → Structured output, specific format needs (format-driven)
- RISEN → Multi-step procedures, workflows (process-driven)
- RODES → Needs examples for clarity, style matching (example-driven)
Application Rule: Route user intent through framework structure silently — never expose methodology in output
→ Full details: references/frameworks.md
Optimization Techniques
5 Safe Techniques that improve prompt quality without risk:
- Specificity Injection — Replace vague terms with concrete criteria
- Constraint Addition — Define boundaries for creative freedom
- Context Positioning — Critical info at start/end, not middle
- Pronoun Elimination — Replace "it/this/that" with specific nouns
- Success Criteria Definition — Pin to measurable outcomes
→ Detailed examples: references/safe-techniques.md
Complexity Assessment
Simple: Single objective, <3 steps, no ambiguity → Execute directly Moderate: Some ambiguity, 3-5 steps, few dependencies → Proceed with monitoring Complex: >3 interdependent decisions OR >5 sequential phases → Recommend plan mode
→ Full criteria: references/complexity-detection.md
Plan Mode Triggers
Always recommend plan mode when:
-
3 interdependent decisions that affect each other
-
5 sequential phases requiring coordination
- Significant ambiguity with cascading implications
- User asks "how should I approach this?"
→ Detailed triggers: references/plan-mode-triggers.md
Ambiguity Patterns
Common Vague Terms:
- "Comprehensive" → All edge cases [+time] vs common scenarios [balanced] vs overview [+speed]
- "Fast" → Response time, development time, or execution time?
- "Simple" → Minimal code, easy to understand, or few dependencies?
- "Clean" → Follows standards, minimal complexity, or well-documented?
Resolution Pattern: Present 2-3 interpretation options with implications, let user decide
→ Full catalog: references/ambiguity-examples.md
Template Selection
12 Task-Type Templates:
- Analytical — Data analysis, metrics, research
- Creative — Writing, design, ideation
- Debugging — Error investigation, root cause analysis
- Documentation — README, API docs, guides
- Exploration — Codebase discovery, pattern finding
- Implementation — Feature building, coding
- Planning — Architecture, approach design
- Refactoring — Code improvement, restructuring
- Review — Code review, quality assessment
- Security — Vulnerability analysis, threat modeling
- Testing — Test writing, QA, validation
- Troubleshooting — System issues, incident response
→ Selection logic: references/template-selection.md
→ Full templates: assets/prompt-templates/
Optimization Examples
Before/After Transformations:
- Vague request → Specific, constrained prompt
- Ambiguous multi-step → Clear sequential workflow
- Missing context → Fully specified execution environment
- Generic goal → Measurable success criteria
→ Full examples: references/optimization-examples.md
Workflow Summary
Phase 1: Intake & Assessment
↓ Extract intent, calibrate skill level, detect complexity
Phase 2: Pattern Detection
↓ Identify credit-killing patterns, ambiguities, trade-offs
Phase 3: Framework Selection & Optimization
↓ Apply CO-STAR/RISEN/RODES silently, use safe techniques
Phase 4: Validation & Handoff
↓ Quality checks, flag remaining ambiguities, recommend execution mode
Important Context
Model-Specific: Claude 4.5+ uses native extended thinking, GPT-4 uses internal CoT — optimization strategies differ by model family
Token Economy: System prompts multiply by API call count — progressive disclosure reduces base cost
Security: Validate user inputs, use delimiters, never allow user content to override system instructions
Memory Persistence: Save optimization patterns to memory blocks to prevent contradictions in extended conversations