name: rebuttal description: "Workflow 4: Submission rebuttal pipeline. Parses external reviews, enforces coverage and grounding, drafts a safe text-only rebuttal under venue limits, and manages follow-up rounds. Use when user says "rebuttal", "reply to reviewers", "ICML rebuttal", "OpenReview response", or wants to answer external reviews safely." argument-hint: [paper-path-or-review-bundle] allowed-tools: Bash(*), Read, Grep, Glob, Write, Edit, Agent, Skill, mcp__codex__codex, mcp__codex__codex-reply
Workflow 4: Rebuttal
Prepare and maintain a grounded, venue-compliant rebuttal for: $ARGUMENTS
Scope
This skill is optimized for:
- text-only rebuttal under strict character/word limits (e.g. ICML single-document)
- per-reviewer thread responses where each reviewer renders independently (e.g. OpenReview-style)
- multiple reviewers with shared and reviewer-specific concerns
- follow-up rounds after the initial rebuttal
- safe drafting with no fabrication, no overpromise, and full issue coverage
This skill does not:
- run new experiments automatically
- generate new theorem claims automatically
- edit or upload a revised PDF
- submit to OpenReview / CMT / HotCRP
If the user already has new results, derivations, or approved commitments, the skill can incorporate them as user-confirmed evidence.
Lifecycle Position
Workflow 1: idea-discovery
Workflow 1.5: experiment-bridge
Workflow 2: auto-review-loop (pre-submission)
Workflow 3: paper-writing
Workflow 4: rebuttal (post-submission external reviews)
Constants
- VENUE =
ICML— Default venue. Override if needed. - RESPONSE_MODE =
TEXT_ONLY— v1 default. - REVIEWER_MODEL =
gpt-5.4— Used via Codex MCP for internal stress-testing. - REVIEWER_BACKEND =
codex— Default: Codex MCP (xhigh). Override with— reviewer: oracle-profor GPT-5.4 Pro via Oracle MCP. Seeshared-references/reviewer-routing.md. - MAX_INTERNAL_DRAFT_ROUNDS = 2 — draft → lint → revise.
- VENUE_MODE =
single_document—single_documentfor one shared author response, orper_reviewer_threadwhen each reviewer thread renders independently. Confirm the venue/interface before drafting if unclear. Affects Phase 4/7 output shape. - STRESS_TEST_ROUNDS_BASE = 1 — One Codex MCP critique round on the full response set. Add focused rounds for
reviewer_priority: pivotalresponses, terminating when Codex returns no new substantive issues. Hard cap at 5. - MAX_FOLLOWUP_ROUNDS = 3 — per reviewer thread.
- AUTO_EXPERIMENT = false — When
true, automatically invoke/experiment-bridgeto run supplementary experiments when the strategy plan identifies reviewer concerns that require new empirical evidence. Whenfalse(default), pause and present the evidence gap to the user for manual handling. - QUICK_MODE = false — When
true, only run Phase 0-3 (parse reviews, atomize concerns, build strategy). OutputsISSUE_BOARD.md+STRATEGY_PLAN.mdand stops — no drafting, no stress test. Useful for quickly understanding what reviewers want before deciding how to respond. - REBUTTAL_DIR =
rebuttal/
Override:
/rebuttal "paper/" — venue: NeurIPS, character limit: 5000
Required Inputs
- Paper source — PDF, LaTeX directory, or narrative summary
- Raw reviews — pasted text, markdown, or PDF with reviewer IDs
- Venue rules — venue name, character/word limit, text-only or revised PDF allowed, rendering mode (one shared response or independent reviewer threads)
- Current stage — initial rebuttal or follow-up round
If venue rules, limit, or rendering mode are missing, stop and ask before drafting.
Safety Model
Three hard gates — if any fails, do NOT finalize:
- Provenance gate — every factual statement maps to:
paper,review,user_confirmed_result,user_confirmed_derivation, orfuture_work. No source = blocked. - Commitment gate — every promise maps to:
already_done,approved_for_rebuttal, orfuture_work_only. Not approved = blocked. - Coverage gate — every reviewer concern ends in:
answered,deferred_intentionally, orneeds_user_input. No issue disappears.
Workflow
Phase 0: Resume or Initialize
- If
rebuttal/REBUTTAL_STATE.mdexists → resume from recorded phase - Otherwise → create
rebuttal/, initialize all output documents - Load paper, reviews, venue rules, any user-confirmed evidence
Phase 1: Validate Inputs and Normalize Reviews
- Validate venue rules are explicit
- Normalize all reviewer text into
rebuttal/REVIEWS_RAW.md(verbatim) - Record metadata in
rebuttal/REBUTTAL_STATE.md - If ambiguous, pause and ask
Phase 2: Atomize and Classify Reviewer Concerns
Create rebuttal/ISSUE_BOARD.md.
For each atomic concern:
issue_id(e.g., R1-C2)reviewer,round,raw_anchor(short quote)issue_type: assumptions / theorem_rigor / novelty / empirical_support / baseline_comparison / complexity / practical_significance / clarity / reproducibility / otherseverity: critical / major / minorreviewer_stance: positive / swing / negative / unknownreviewer_priority: standard / pivotalpivotal— a reviewer whose response is likely to affect the decision if addressed well: low or borderline rating, addressable concerns, and enough confidence/influence to matter. Phase 3 allocates extra drafting and stress-test budget here.
response_mode: direct_clarification / grounded_evidence / nearest_work_delta / assumption_hierarchy / narrow_concession / future_work_boundary / structural_distinctionstructural_distinction— for "your method reduces to X / is just generic Y / is subsumed by Z" attacks. Pattern: agree on the local reduction; show the structural feature your parameterization preserves that X/Y/Z does not capture, backed by a concrete mechanism (theorem dependency, derivation step, or empirical consequence). Never use rhetorically without the supporting mechanism.
status: open / answered / deferred / needs_user_input
Phase 3: Build Strategy Plan
Create rebuttal/STRATEGY_PLAN.md.
- Identify 2-4 global themes resolving shared concerns
- Choose response mode per issue
- Build character budget (10-15% opener, 75-80% per-reviewer, 5-10% closing) — applies in
single_documentmode; inper_reviewer_threadmode, set per-thread word/char targets instead - Identify pivotal reviewer(s) — reviewers whose vote or confidence shift would most affect the decision, especially when concerns are addressable rather than ideological. Mark them
reviewer_priority: pivotalinISSUE_BOARD.md. There may be more than one. Allocate disproportionate drafting + stress-test budget here. - Identify blocked claims (ungrounded or unapproved)
- If unresolved blockers → pause and present to user
QUICK_MODE exit: If QUICK_MODE = true, stop here. Present ISSUE_BOARD.md + STRATEGY_PLAN.md to the user and summarize: how many issues per reviewer, shared vs unique concerns, recommended priorities, and evidence gaps. The user can then decide to continue with full rebuttal (/rebuttal — quick mode: false) or write manually.
Phase 3.5: Evidence Sprint (when AUTO_EXPERIMENT = true)
Skip entirely if AUTO_EXPERIMENT is false — instead, pause and present the evidence gaps to the user.
If the strategy plan identifies issues that require new empirical evidence (tagged response_mode: grounded_evidence with evidence_source: needs_experiment):
-
Generate a mini experiment plan from the reviewer concerns:
- What to run (ablation, baseline comparison, scale-up, condition check)
- Success criterion (what result would satisfy the reviewer)
- Estimated GPU-hours
-
Invoke
/experiment-bridgewith the mini plan:/experiment-bridge "rebuttal/REBUTTAL_EXPERIMENT_PLAN.md" -
Wait for results, then update
ISSUE_BOARD.md:- Tag completed experiments as
user_confirmed_result - Update evidence source for relevant issue cards
- Tag completed experiments as
-
If experiments fail or are inconclusive:
- Switch response mode to
narrow_concessionorfuture_work_boundary - Do NOT fabricate positive results
- Switch response mode to
-
Save experiment results to
rebuttal/REBUTTAL_EXPERIMENTS.mdfor provenance tracking.
Time guard: If estimated GPU-hours exceed rebuttal deadline, skip and flag for manual handling.
Phase 4: Draft Initial Rebuttal
Create the draft artifact(s) per VENUE_MODE:
single_documentmode → onerebuttal/REBUTTAL_DRAFT_v1.mdper_reviewer_threadmode → onerebuttal/Reviewer_<ID>_response.mdper reviewer (no top-levelREBUTTAL_DRAFT_v1.md)
Structure depends on VENUE_MODE:
-
single_document— oneREBUTTAL_DRAFT_v1.md:- Short opener — thank reviewers + 2-4 global resolutions
- Per-reviewer numbered responses — answer → evidence → implication
- Short closing — resolved / remaining / acceptance case
-
per_reviewer_thread— one self-containedReviewer_<ID>_response.mdper reviewer:- Brief acknowledgment of that reviewer's main thrust
- Numbered W#/Q# responses (answer → evidence → implication)
- Optional shared experimental-setup paragraph (see "Reusable setup block" below)
- Each file must be readable standalone. No "see Reviewer X's response" references. No global opener.
Default reply pattern per issue:
- Sentence 1: direct answer
- Sentence 2-4: grounded evidence
- Last sentence: implication for the paper
Reusable setup block (per_reviewer_thread mode).
If multiple reviewer-thread responses need the same experimental setup or metric definitions, write a canonical SETUP_METRICS_BLOCK.md. Reuse it consistently in each reviewer file that needs it. Target ≤ 150 words; expand only with genuinely reviewer-specific additions inline. Change-once-update-everywhere prevents drift across threads.
Heuristics from successful rebuttals (content):
- Evidence > assertion
- Global narrative first, per-reviewer detail second (single_document mode only)
- Concrete numbers for counter-intuitive points
- Name closest prior work + exact delta for novelty disputes
- Concede narrowly when reviewer is right
- For theory: separate core vs technical assumptions
- Answer friendly reviewers too
Reviewer-defensive moves:
- Minimum sufficient evidence per concern. Usually one numerical anchor: the metric that maps directly to that reviewer's specific ask. Cut metrics other reviewers care about — bloat dilutes the answer.
- Pre-registered calibration phrasing. When a threshold or hold-out was fixed before generated samples were inspected, say so explicitly with a phrase like "set on hold-out before any generated sample was inspected." Defuses cherry-pick attacks at near-zero word cost. Only use when actually true.
- Surface non-obvious design choices upfront. If the experimental setup has a non-obvious caveat (compute-matched ≠ epoch-matched, atypical seed protocol, restricted parameter subset, etc.), name it concretely with numbers where they clarify the design choice. Pre-empts adversarial reverse engineering.
- Structural distinction over denial. When a reviewer claims your work reduces to / is subsumed by a generic framework, do not deny the reduction. Identify the structural feature your parameterization preserves that the generic framework does not — see
response_mode: structural_distinction. - Concede without surrendering the claim. When the reviewer is partly right, explicitly accept the local point, then state what remains true and why it still supports the paper's contribution. Pair the concession with the preserved theorem, mechanism, empirical result, or scope condition.
Hard rules:
- NEVER invent experiments, numbers, derivations, citations, or links
- NEVER promise what user hasn't approved
- If no strong evidence exists, say less not more
single_document mode only: also generate rebuttal/PASTE_READY.txt (plain text, exact character count for the OpenReview/CMT paste field). In per_reviewer_thread mode skip this artifact — each Reviewer_<ID>_response.md is itself the paste target for its thread.
Also generate rebuttal/REVISION_PLAN.md — the overall revision checklist.
This document is the single source of truth for every paper revision promised (explicitly or implicitly) in the rebuttal draft. It exists so the author can track follow-through after the rebuttal is submitted, and so the commitment gate in Phase 5 has a concrete artifact to validate against.
Structure:
-
Header
- Paper title, venue, character limit, rebuttal round
- Links back to
ISSUE_BOARD.md,STRATEGY_PLAN.md,REBUTTAL_DRAFT_v1.md
-
Overall checklist — a single flat GitHub-style checklist covering every revision item, so the author can tick items off as they land in the camera-ready / revised PDF:
## Overall Checklist - [ ] (R1-C2) Add assumption hierarchy table to Section 3.1 — commitment: `approved_for_rebuttal` — owner: author — status: pending - [ ] (R2-C1) Clarify novelty delta vs. Smith'24 in Section 2 related work — commitment: `already_done` — status: verify wording - [ ] (R3-C4) Add runtime breakdown figure to Appendix B — commitment: `future_work_only` — status: deferred, note in camera-ready - ...Checklist items must be atomic (one paper edit per line) and each must reference its
issue_idso it maps back toISSUE_BOARD.md. -
Grouped view — the same items regrouped by (a) paper section/location and (b) severity, so the author can plan the revision pass efficiently.
-
Commitment summary — counts of
already_done/approved_for_rebuttal/future_work_only, plus anyneeds_user_inputitems that are blocking. -
Out-of-scope log — reviewer concerns that will not trigger a paper revision (e.g.
deferred_intentionally,narrow_concessionwith no edit), with a one-line reason each. This keeps the checklist honest: nothing silently disappears.
Rules for REVISION_PLAN.md:
- Every checklist item must map to at least one
issue_idfromISSUE_BOARD.md. - Every promise in
REBUTTAL_DRAFT_v1.mdthat implies a paper edit must appear as a checklist item — if it is not in the plan, it is a commitment-gate violation. - Never add items that are not backed by the draft or by user-confirmed evidence.
- On rerun / follow-up rounds, update checkbox state in place rather than regenerating from scratch.
Phase 5: Safety Validation
Run all lints:
- Coverage — every issue maps to draft anchor
- Provenance — every factual sentence has source
- Commitment — promises are approved AND every paper-edit promise in the draft appears as a checklist item in
REVISION_PLAN.md(and vice versa — no orphan items in the plan) - Tone — flag aggressive/submissive/evasive phrases
- Consistency — no contradictions across reviewer replies
- Limit — exact character count, compress if over (redundancy → friendly → opener → wording, never drop critical answers)
- Thread-local context (
per_reviewer_threadmode only) — each reviewer file must be intelligible without reading any other reviewer file. Flag any "see Reviewer X" references or undefined terms that rely on cross-thread context. - Adversarial design-choice scan — for each experimental claim, ask: "Could a hostile reviewer find a non-obvious design choice (compute-match, frozen subset, sampling protocol) that I haven't disclosed?" If yes, add a one-line caveat in the Setup paragraph. Narrower than provenance; focused on design choices not factual sources.
Phase 6: Codex MCP Stress Test
mcp__codex__codex:
config: {"model_reasoning_effort": "xhigh"}
prompt: |
Stress-test this rebuttal draft:
[raw reviews + issue board + draft + venue rules]
1. Unanswered or weakly answered concerns?
2. Unsupported factual statements?
3. Risky or unapproved promises?
4. Tone problems?
5. Paragraph most likely to backfire with meta-reviewer?
6. Minimal grounded fixes only. Do NOT invent evidence.
Verdict: safe to submit / needs revision
Iterations. Run the base round on the full draft. Then run focused follow-up rounds on each reviewer_priority: pivotal response, terminating when Codex returns no new substantive issues. Hard cap at 5 rounds total. Save each round to rebuttal/MCP_STRESS_TEST_round<N>.md; the highest round number represents the final state. If any hard safety blocker remains → revise before finalizing.
Phase 7: Finalize
Outputs depend on VENUE_MODE:
single_document mode — produce two versions:
-
rebuttal/PASTE_READY.txt— strict version- Plain text, exact character count, fits venue limit
- Ready to paste directly into the venue interface
- No markdown formatting, no extras
-
rebuttal/REBUTTAL_DRAFT_rich.md— extended version- Same structure but with more detail: fuller explanations, additional evidence, optional paragraphs
- Marked with
[OPTIONAL — cut if over limit]for sections that exceed the strict version - Author can read this to understand the full reasoning, then manually decide what to keep/cut/rewrite
- Useful for follow-up rounds — the extra material is pre-written
per_reviewer_thread mode — produce one file per reviewer:
rebuttal/Reviewer_<ID>_response.md— one self-contained file per reviewer, ready to paste into the corresponding reviewer threadrebuttal/SETUP_METRICS_BLOCK.md— optional canonical setup/metrics text when reused across reviewer filesrebuttal/SUPPLEMENTARY_FIG_PDF/(optional) — when the venue does not allow PDF revision but allows anonymous figure links, generate a venue-compliant supplementary PDF. Do not hard-code an anonymous-hosting platform or typesetting style; choose what the target venue accepts.
Both modes:
- Update
rebuttal/REBUTTAL_STATE.md - Refresh
rebuttal/REVISION_PLAN.mdso the overall checklist matches the final draft (add items, markalready_doneas checked, carry forward anypendingitems) - Present to user:
- For
single_document:PASTE_READY.txtcharacter count vs venue limit;REBUTTAL_DRAFT_rich.mdfor review - For
per_reviewer_thread: list of per-reviewer files with word counts;SETUP_METRICS_BLOCK.mdif used; supplementary PDF if generated REVISION_PLAN.mdchecklist — counts of pending / approved / deferred- Remaining risks + lines needing manual approval
- For
Phase 8: Follow-Up Rounds
When new reviewer comments arrive:
- Append verbatim to
rebuttal/FOLLOWUP_LOG.md - Link to existing issues or create new ones
- Draft delta reply only (not full rewrite)
- Update
rebuttal/REVISION_PLAN.mdin place — add any new checklist items introduced by the follow-up, tick off items the author has already completed, and keep existing items' status current - Re-run safety lints
- Use Codex MCP reply for continuity if useful
- Rules: escalate technically not rhetorically; concede if reviewer is correct; stop arguing if reviewer is immovable and no new evidence exists
Key Rules
- Large file handling: If Write fails, retry with Bash heredoc silently.
- Never fabricate. No invented evidence, numbers, derivations, citations, or links.
- Never overpromise. Only promise what user explicitly approved.
- Full coverage. Every reviewer concern tracked and accounted for.
- Preserve raw records. Reviews and MCP outputs stored verbatim.
- Structure follows
VENUE_MODE.single_document→ global opener with shared concerns + per-reviewer numbered detail.per_reviewer_thread→ no global opener; each reviewer file is self-contained. - Answer friendly reviewers too. Reinforce supportive framing.
- Meta-reviewer closing. Summarize resolved/remaining/why accept.
- Evidence > rhetoric. Derivations and numbers over prose.
- Concede selectively. Narrow honest concessions > broad denials.
- Don't waste space on unwinnable arguments. Answer once, move on.
- Respect the limit. Character budget is a hard constraint.
- Resume cleanly. Continue from REBUTTAL_STATE.md on rerun.
- Anti-hallucination citations. Any reference added must go through DBLP → CrossRef → [VERIFY].
Review Tracing
After each mcp__codex__codex or mcp__codex__codex-reply reviewer call, save the trace following shared-references/review-tracing.md. Use tools/save_trace.sh or write files directly to .aris/traces/<skill>/<date>_run<NN>/. Respect the --- trace: parameter (default: full).