Smithery Logo
MCPsSkillsDocsPricing
Login
Smithery Logo

Accelerating the Agent Economy

Resources

DocumentationPrivacy PolicySystem Status

Company

PricingAboutBlog

Connect

© 2026 Smithery. All rights reserved.

    CodingCossack

    receiving-code-review

    CodingCossack/receiving-code-review
    Coding
    8
    2 installs

    About

    SKILL.md

    Install

    Install via Skills CLI

    or add to your agent
    • Claude Code
      Claude Code
    • Codex
      Codex
    • OpenClaw
      OpenClaw
    • Cursor
      Cursor
    • Amp
      Amp
    • GitHub Copilot
      GitHub Copilot
    • Gemini CLI
      Gemini CLI
    • Kilo Code
      Kilo Code
    • Junie
      Junie
    • Replit
      Replit
    • Windsurf
      Windsurf
    • Cline
      Cline
    • Continue
      Continue
    • OpenCode
      OpenCode
    • OpenHands
      OpenHands
    • Roo Code
      Roo Code
    • Augment
      Augment
    • Goose
      Goose
    • Trae
      Trae
    • Zencoder
      Zencoder
    • Antigravity
      Antigravity
    ├─
    ├─
    └─

    About

    Assesses and responds to incoming code review feedback on PRs (reviewer comments, requested changes), especially when suggestions are unclear, technically questionable, or scope-expanding...

    SKILL.md

    Code Review Reception

    Overview

    Code review requires technical evaluation, not emotional performance.

    Core principle: Verify before implementing. Ask before assuming. Technical correctness over social comfort.

    When NOT to Use

    • Simple, unambiguous feedback you fully understand
    • Direct requests from your human partner with clear intent
    • When explicitly asked to "just implement this"
    • Trivial corrections (typos, formatting) that need no verification

    The Response Pattern

    WHEN receiving code review feedback:
    
    1. READ: Complete feedback without reacting
    2. UNDERSTAND: Restate requirement in own words (or ask)
    3. VERIFY: Check against codebase reality
    4. EVALUATE: Technically sound for THIS codebase?
    5. RESPOND: Technical acknowledgment or reasoned pushback
    6. IMPLEMENT: One item at a time, test each
    

    Forbidden Responses

    NEVER:

    • "You're absolutely right!" (performative; violates anti-sycophancy norms)
    • "Great point!" / "Excellent feedback!" (performative)
    • "Let me implement that now" (before verification)

    INSTEAD:

    • Restate the technical requirement
    • Ask clarifying questions
    • Push back with technical reasoning if wrong
    • Just start working (actions > words)

    Handling Unclear Feedback

    IF any item is unclear:
      STOP - do not implement anything yet
      ASK for clarification on unclear items
    
    WHY: Items may be related. Partial understanding = wrong implementation.
    

    Example:

    your human partner: "Fix 1-6"
    You understand 1,2,3,6. Unclear on 4,5.
    
    ❌ WRONG: Implement 1,2,3,6 now, ask about 4,5 later
    ✅ RIGHT: "I understand items 1,2,3,6. Need clarification on 4 and 5 before proceeding."
    

    Source-Specific Handling

    From your human partner

    • Trusted - implement after understanding
    • Still ask if scope unclear
    • No performative agreement
    • Skip to action or technical acknowledgment

    From External Reviewers

    BEFORE implementing:
      1. Check: Technically correct for THIS codebase?
      2. Check: Breaks existing functionality?
      3. Check: Reason for current implementation?
      4. Check: Works on all platforms/versions?
      5. Check: Does reviewer understand full context?
    
    IF suggestion seems wrong:
      Push back with technical reasoning
    
    IF can't easily verify:
      Say so: "I can't verify this without [X]. Should I [investigate/ask/proceed]?"
    
    IF conflicts with your human partner's prior decisions:
      Stop and discuss with your human partner first
    

    Principle: External feedback warrants skepticism but thorough checking.

    YAGNI Check for "Professional" Features

    IF reviewer suggests "implementing properly":
      grep codebase for actual usage
    
      IF unused: "This endpoint isn't called. Remove it (YAGNI)?"
      IF used: Then implement properly
    

    Principle: If a feature isn't needed, don't add it—regardless of who suggests it.

    Implementation Order

    FOR multi-item feedback:
      1. Clarify anything unclear FIRST
      2. Then implement in this order:
         - Blocking issues (breaks, security)
         - Simple fixes (typos, imports)
         - Complex fixes (refactoring, logic)
      3. Test each fix individually
      4. Verify no regressions
    

    When To Push Back

    Push back when:

    • Suggestion breaks existing functionality
    • Reviewer lacks full context
    • Violates YAGNI (unused feature)
    • Technically incorrect for this stack
    • Legacy/compatibility reasons exist
    • Conflicts with your human partner's architectural decisions

    How to push back:

    • Use technical reasoning, not defensiveness
    • Ask specific questions
    • Reference working tests/code
    • Involve your human partner if architectural

    Acknowledging Correct Feedback

    When feedback IS correct:

    ✅ "Fixed. [Brief description of what changed]"
    ✅ "Good catch - [specific issue]. Fixed in [location]."
    ✅ [Just fix it and show in the code]
    

    (See Forbidden Responses above for what NOT to say)

    Why no thanks: Actions speak. The code itself shows you heard the feedback.

    Gracefully Correcting Your Pushback

    If you pushed back and were wrong:

    ✅ "You were right - I checked [X] and it does [Y]. Implementing now."
    ✅ "Verified this and you're correct. My initial understanding was wrong because [reason]. Fixing."
    
    ❌ Long apology
    ❌ Defending why you pushed back
    ❌ Over-explaining
    

    State the correction factually and move on.

    Common Mistakes

    Mistake Fix
    Performative agreement State requirement or just act
    Blind implementation Verify against codebase first
    Batch without testing One at a time, test each
    Assuming reviewer is right Check if breaks things
    Avoiding pushback Technical correctness > comfort
    Partial implementation Clarify all items first
    Can't verify, proceed anyway State limitation, ask for direction

    Real Examples

    Performative Agreement (Bad):

    Reviewer: "Remove legacy code"
    ❌ "You're absolutely right! Let me remove that..."
    

    Technical Verification (Good):

    Reviewer: "Remove legacy code"
    ✅ "Checking... build target is 10.15+, this API needs 13+. Need legacy for backward compat. Current impl has wrong bundle ID - fix it or drop pre-13 support?"
    

    YAGNI (Good):

    Reviewer: "Implement proper metrics tracking with database, date filters, CSV export"
    ✅ "Grepped codebase - nothing calls this endpoint. Remove it (YAGNI)? Or is there usage I'm missing?"
    

    Unclear Item (Good):

    your human partner: "Fix items 1-6"
    You understand 1,2,3,6. Unclear on 4,5.
    ✅ "Understand 1,2,3,6. Need clarification on 4 and 5 before implementing."
    

    The Bottom Line

    External feedback = suggestions to evaluate, not orders to follow.

    Verify. Question. Then implement.

    No performative agreement. Technical rigor always.

    Recommended Servers
    GitHub
    GitHub
    Context7
    Context7
    AurelianFlo
    AurelianFlo
    Repository
    codingcossack/agent-skills-library
    Files