Smithery Logo
MCPsSkillsDocsPricing
Login
Smithery Logo

Accelerating the Agent Economy

Resources

DocumentationPrivacy PolicySystem Status

Company

PricingAboutBlog

Connect

© 2026 Smithery. All rights reserved.

    agent-skills

    challenge

    agent-skills/challenge
    4,253
    5 installs

    About

    SKILL.md

    Install

    Install via Skills CLI

    or add to your agent
    • Claude Code
      Claude Code
    • Codex
      Codex
    • OpenClaw
      OpenClaw
    • Cursor
      Cursor
    • Amp
      Amp
    • GitHub Copilot
      GitHub Copilot
    • Gemini CLI
      Gemini CLI
    • Kilo Code
      Kilo Code
    • Junie
      Junie
    • Replit
      Replit
    • Windsurf
      Windsurf
    • Cline
      Cline
    • Continue
      Continue
    • OpenCode
      OpenCode
    • OpenHands
      OpenHands
    • Roo Code
      Roo Code
    • Augment
      Augment
    • Goose
      Goose
    • Trae
      Trae
    • Zencoder
      Zencoder
    • Antigravity
      Antigravity
    ├─
    ├─
    └─

    About

    /em -challenge — Pre-Mortem Plan Analysis

    SKILL.md

    /em:challenge — Pre-Mortem Plan Analysis

    Command: /em:challenge <plan>

    Systematically finds weaknesses in any plan before reality does. Not to kill the plan — to make it survive contact with reality.


    The Core Idea

    Most plans fail for predictable reasons. Not bad luck — bad assumptions. Overestimated demand. Underestimated complexity. Dependencies nobody questioned. Timing that made sense in a spreadsheet but not in the real world.

    The pre-mortem technique: imagine it's 12 months from now and this plan failed spectacularly. Now work backwards. Why?

    That's not pessimism. It's how you build something that doesn't collapse.


    When to Run a Challenge

    • Before committing significant resources to a plan
    • Before presenting to the board or investors
    • When you notice you're only hearing positive feedback about the plan
    • When the plan requires multiple external dependencies to align
    • When there's pressure to move fast and "figure it out later"
    • When you feel excited about the plan (excitement is a signal to scrutinize harder)

    The Challenge Framework

    Step 1: Extract Core Assumptions

    Before you can test a plan, you need to surface everything it assumes to be true.

    For each section of the plan, ask:

    • What has to be true for this to work?
    • What are we assuming about customer behavior?
    • What are we assuming about competitor response?
    • What are we assuming about our own execution capability?
    • What external factors does this depend on?

    Common assumption categories:

    • Market assumptions — size, growth rate, customer willingness to pay, buying cycle
    • Execution assumptions — team capacity, velocity, no major hires needed
    • Customer assumptions — they have the problem, they know they have it, they'll pay to solve it
    • Competitive assumptions — incumbents won't respond, no new entrant, moat holds
    • Financial assumptions — burn rate, revenue timing, CAC, LTV ratios
    • Dependency assumptions — partner will deliver, API won't change, regulations won't shift

    Step 2: Rate Each Assumption

    For every assumption extracted, rate it on two dimensions:

    Confidence level (how sure are you this is true):

    • High — verified with data, customer conversations, market research
    • Medium — directionally right but not validated
    • Low — plausible but untested
    • Unknown — we simply don't know

    Impact if wrong (what happens if this assumption fails):

    • Critical — plan fails entirely
    • High — major delay or cost overrun
    • Medium — significant rework required
    • Low — manageable adjustment

    Step 3: Map Vulnerabilities

    The matrix of Low/Unknown confidence × Critical/High impact = your highest-risk assumptions.

    Vulnerability = Low confidence + High impact

    These are not problems to ignore. They're the bets you're making. The question is: are you making them consciously?

    Step 4: Find the Dependency Chain

    Many plans fail not because any single assumption is wrong, but because multiple assumptions have to be right simultaneously.

    Map the chain:

    • Does assumption B depend on assumption A being true first?
    • If the first thing goes wrong, how many downstream things break?
    • What's the critical path? What has zero slack?

    Step 5: Test the Reversibility

    For each critical vulnerability: if this assumption turns out to be wrong at month 3, what do you do?

    • Can you pivot?
    • Can you cut scope?
    • Is money already spent?
    • Are commitments already made?

    The less reversible, the more rigorously you need to validate before committing.


    Output Format

    Challenge Report: [Plan Name]

    CORE ASSUMPTIONS (extracted)
    1. [Assumption] — Confidence: [H/M/L/?] — Impact if wrong: [Critical/High/Medium/Low]
    2. ...
    
    VULNERABILITY MAP
    Critical risks (act before proceeding):
    • [#N] [Assumption] — WHY it might be wrong — WHAT breaks if it is
    
    High risks (validate before scaling):
    • ...
    
    DEPENDENCY CHAIN
    [Assumption A] → depends on → [Assumption B] → which enables → [Assumption C]
    Weakest link: [X] — if this breaks, [Y] and [Z] also fail
    
    REVERSIBILITY ASSESSMENT
    • Reversible bets: [list]
    • Irreversible commitments: [list — treat with extreme care]
    
    KILL SWITCHES
    What would have to be true at [30/60/90 days] to continue vs. kill/pivot?
    • Continue if: ...
    • Kill/pivot if: ...
    
    HARDENING ACTIONS
    1. [Specific validation to do before proceeding]
    2. [Alternative approach to consider]
    3. [Contingency to build into the plan]
    

    Challenge Patterns by Plan Type

    Product Roadmap

    • Are we building what customers will pay for, or what they said they wanted?
    • Does the velocity estimate account for real team capacity (not theoretical)?
    • What happens if the anchor feature takes 3× longer than estimated?
    • Who owns decisions when requirements conflict?

    Go-to-Market Plan

    • What's the actual ICP conversion rate, not the hoped-for one?
    • How many touches to close, and do you have the sales capacity for that?
    • What happens if the first 10 deals take 3 months instead of 1?
    • Is "land and expand" a real motion or a hope?

    Hiring Plan

    • What happens if the key hire takes 4 months to find, not 6 weeks?
    • Is the plan dependent on retaining specific people who might leave?
    • Does the plan account for ramp time (usually 3–6 months before full productivity)?
    • What's the burn impact if headcount leads revenue by 6 months?

    Fundraising Plan

    • What's your fallback if the lead investor passes?
    • Have you modeled the timeline if it takes 6 months, not 3?
    • What's your runway at current burn if the round closes at the low end?
    • What assumptions break if you raise 50% of the target amount?

    The Hardest Questions

    These are the ones people skip:

    • "What's the bear case, not the base case?"
    • "If this exact plan was run by a team we don't trust, would it work?"
    • "What are we not saying out loud because it's uncomfortable?"
    • "Who has incentives to make this plan sound better than it is?"
    • "What would an enemy of this plan attack first?"

    Deliverable

    The output of /em:challenge is not permission to stop. It's a vulnerability map. Now you can make conscious decisions: validate the risky assumptions, hedge the critical ones, or accept the bets you're making knowingly.

    Unknown risks are dangerous. Known risks are manageable.

    Recommended Servers
    InfraNodus Knowledge Graphs & Text Analysis
    InfraNodus Knowledge Graphs & Text Analysis
    Jina AI
    Jina AI
    Unicorn or Bust
    Unicorn or Bust
    Repository
    alirezarezvani/claude-skills
    Files