name: receive-feedback description: Process external code review feedback with technical rigor. Use when receiving feedback from another LLM, human reviewer, or CI tool. Verifies claims before implementing, tracks disposition. disable-model-invocation: false
Receive Feedback
Overview
Process code review feedback with verification-first discipline. No performative agreement. Technical correctness over social comfort.
Quick Reference
┌─────────────┐ ┌──────────────┐ ┌─────────────┐
│ VERIFY │ ──▶ │ EVALUATE │ ──▶ │ EXECUTE │
│ (tool-based)│ │ (decision │ │ (implement/ │
│ │ │ matrix) │ │ reject/ │
└─────────────┘ └──────────────┘ │ defer) │
└─────────────┘
Core Principle
Verify before implementing. Ask before assuming.
When To Use
- Receiving code review from another LLM session
- Processing PR review comments
- Evaluating CI/linter feedback
- Handling suggestions from pair programming
Workflow
For each feedback item:
- Verify - Use tools to check if feedback is technically valid
- Evaluate - Apply decision matrix to determine action
- Execute - Implement, reject with evidence, or defer
Hard gates
Do not advance to the next gate until its pass condition is true. Details live in VERIFICATION.md, EVALUATION.md, and RESPONSE.md.
Gate 1 — Verification evidence (per item)
- Locate the referenced code or behavior (
Read,Grep, tests, or another check fromVERIFICATION.mdthat fits the claim). - Record the outcome: claim holds, fails, or unclear, with an artifact (file path and line range, command/test result, or short quoted tool output—not paraphrase alone).
Pass when: That record exists before you choose implement / reject / defer / clarify. Fail (stop): Proceeding with only a verbal “I verified” without an artifact.
Gate 2 — Disposition locked (per item)
- Apply the decision matrix in
EVALUATION.mdto the verified facts. - Choose exactly one primary outcome: implement, reject with evidence, defer, or ask for clarification.
Pass when: Outcome is explicit. Fail (stop): Implementing while the item is still ambiguous—resolve or clarify first per EVALUATION.md.
Gate 3 — Response artifact (batch)
- After every item has passed Gates 1–2, fill the structured template in
RESPONSE.md. - Ensure rejections include verification evidence and implemented rows cite locations.
Pass when: Each feedback item appears in the correct section of the response template with required columns satisfied. Fail (stop): Shipping a summary that omits an item or lacks evidence on a rejection.
Command Workflow
Use this skill from the /receive-feedback command or by invoking it directly with a feedback file path.
- Read the feedback file at
$ARGUMENTS - Parse individual feedback items, whether numbered, bulleted, or freeform
- Load this skill:
Skill(skill: "beagle-core:receive-feedback") - Process each item through verify → evaluate → execute
- Produce the structured response summary defined in
RESPONSE.md
Expected Feedback File Format
The feedback file may contain numbered or bulleted items:
1. Remove unused import on line 15
2. Add error handling to the API call
3. Consider using a generator for large datasets
4. Fix typo in variable name: `usr` → `user`
Freeform prose is also acceptable; extract actionable items from the text.
Example
/receive-feedback reviews/pr-123-feedback.md
Reads the file, processes each item with technical verification, and outputs a structured response table.
Files
VERIFICATION.md- Tool-based verification workflowEVALUATION.md- Decision matrix and rulesRESPONSE.md- Structured output formatreferences/skill-integration.md- Using with code-review skills