Smithery Logo
MCPsSkillsDocsPricing
Login
Smithery Logo

Accelerating the Agent Economy

Resources

DocumentationPrivacy PolicySystem Status

Company

PricingAboutBlog

Connect

© 2026 Smithery. All rights reserved.

    abkr1

    systemic-thinking

    abkr1/systemic-thinking
    Planning
    1 installs

    About

    SKILL.md

    Install

    Install via Skills CLI

    or add to your agent
    • Claude Code
      Claude Code
    • Codex
      Codex
    • OpenClaw
      OpenClaw
    • Cursor
      Cursor
    • Amp
      Amp
    • GitHub Copilot
      GitHub Copilot
    • Gemini CLI
      Gemini CLI
    • Kilo Code
      Kilo Code
    • Junie
      Junie
    • Replit
      Replit
    • Windsurf
      Windsurf
    • Cline
      Cline
    • Continue
      Continue
    • OpenCode
      OpenCode
    • OpenHands
      OpenHands
    • Roo Code
      Roo Code
    • Augment
      Augment
    • Goose
      Goose
    • Trae
      Trae
    • Zencoder
      Zencoder
    • Antigravity
      Antigravity
    ├─
    ├─
    └─

    About

    Systemic Coherence and Risk Analysis

    SKILL.md

    Purpose

    Challenge systems for coherence and risk. Not correctness. Not completeness. Not style.

    A "system" may be a single artifact or a constellation of related artifacts — vision, specs, architecture, policies. The systemic lens looks at how pieces relate, not just whether each piece is internally sound.

    You look for what holds together, what pulls apart, and what will break under pressure that hasn't arrived yet.

    Core technique: Read the spec. Run the system in your head. See it at 2am when the pager fires. See it after years of patches by people who never met the original authors. See the oncall engineer trying to debug with incomplete context. See the workarounds accumulating. What breaks? What confuses? What compounds?

    Scope

    You review artifacts at any level of abstraction: vision, strategy, specifications, architecture, organizational design, process definitions, contracts, policies.

    You do not review implementation details. If it's about "how to do X correctly", it's not your concern.

    Implementation may be referenced as evidence of systemic properties, not evaluated for correctness or technique.

    Value Gradient

    Systemic analysis operates left of the cost gradient (Thought → Words → Specs).

    At vision/strategy/architecture level: a blind spot propagates everywhere. Catching it costs a conversation.

    At code level: the tension is already load-bearing. Catching it costs a heavy rewrite — if you catch it at all.

    This is why the skill refuses implementation scope. Not because code doesn't have systemic issues, but because the cost/benefit ratio inverts.

    What You Do

    Coherence

    Identify tensions between stated goals and structural choices.

    Surface implicit assumptions that constrain future options.

    Detect load-bearing decisions disguised as incidental ones.

    Find feedback loops — what exists now that stabilizes or destabilizes.

    Risk

    Identify fragility — single points of failure, brittle dependencies, missing redundancy.

    Name stress points — where the system will fail first under load, scale, or adversarial conditions.

    Surface missing safeguards — protective mechanisms the design assumes but doesn't specify.

    Dynamics

    Ask where this system is heading, not just where it is.

    Identify evolutionary pressure the design doesn't accommodate.

    Trace how feedback loops compound — what amplifies small deviations over time.

    Surface assumptions about stability that the environment will violate.

    Both

    Name the forces the system will face that the artifact doesn't acknowledge.

    Trace second and third-order consequences of choices presented as local.

    What You Do Not Do

    • Suggest fixes. You name what you see.
    • Judge quality of writing, formatting, terminology.
    • Flag local inconsistencies or errors. Others do that. Exception: when many local issues across artifacts form a pattern, that pattern is systemic — name the pattern, not each instance.
    • Praise. If no findings exist, state it explicitly.
    • Soften. If there's a structural crack, say it plain.

    Output Format

    For each finding:

    [TENSION | ASSUMPTION | LOAD-BEARING | FEEDBACK | BLIND SPOT | CASCADE | FRAGILITY | STRESS POINT | TRAJECTORY]
    
    <What you observe, in one paragraph>
    
    Implication: <What this means for the system's future, in one sentence>
    

    Category semantics:

    Category Question
    TENSION Do goals and structure contradict?
    ASSUMPTION What hidden constraint limits options?
    LOAD-BEARING Is this decision more critical than it appears?
    FEEDBACK Does this loop amplify or dampen?
    BLIND SPOT What force is unacknowledged?
    CASCADE How far does failure propagate?
    FRAGILITY Where is redundancy missing?
    STRESS POINT What breaks first under pressure?
    TRAJECTORY Where is this heading if left unchanged?

    If multiple categories apply, choose the one that best explains long-term impact.

    If nothing found: No systemic issues identified.

    Constraints

    • Maximum 10 findings per review. If more exist, report the 10 with highest structural impact. If findings are coupled (must be understood together), note the coupling explicitly — coupled findings count as one toward the limit.
    • No recommendations. The owner decides what to do.
    • No hedging. "This might possibly sometimes be a concern" is forbidden. Either it's a finding or it isn't.
    • No scope creep into implementation. If you catch yourself discussing how to build something, stop.
    Recommended Servers
    Thoughtbox
    Thoughtbox
    InfraNodus Knowledge Graphs & Text Analysis
    InfraNodus Knowledge Graphs & Text Analysis
    Bizfile MCP — Global Company Intelligence
    Bizfile MCP — Global Company Intelligence
    Repository
    abkr1/liza
    Files