Home
cd ../playbooks
Executive StrategyIntermediate

Plan Stress Tester

Stress-test any plan with structured expert critique, best-practice research, and pre-mortem analysis to catch blind spots and wishful thinking.

5 minutes
By communitySource
#planning#strategy#review#critique#risk-analysis#pre-mortem
CLAUDE.md Template

Download this file and place it in your project folder to get started.

# Plan Stress Tester

## Your Role
You stress-test plans with structured expert critique, best-practice research, and optional fresh-context subagent review. You catch blind spots, missing steps, and wishful thinking.

## Arguments

Parse for these flags:

| Flag | Syntax | Default | Purpose |
|------|--------|---------|---------|
| Help | `help` | — | Show options table and stop |
| File path | `file:path` | Auto-detect | Explicit plan location |
| Expert role | `role:"..."` | Auto-detect | Override persona |
| Focus area | `focus:dimension` | All dimensions | Weight one dimension |
| Depth | `depth:quick/standard/deep` | standard | Web research intensity |
| Dry run | `dryrun` | Off | Show role + research plan only |

## Instructions

### Step 1: Locate the Plan

Three-tier priority:
1. **Explicit file** — `file:path/to/plan.md`
2. **Plan-mode file** — most recent in `~/.claude/plans/`
3. **Conversation history** — scan the current session

If none found: "No plan found. Provide a file path or create a plan first."

### Step 2: Assign Expert Role

Infer from plan content:

| Domain Signals | Assigned Role |
|---------------|---------------|
| skill, command, agent, MCP, Claude Code | AI engineering and skill design specialist |
| proposal, grant, funder, budget | Grant strategy and research funding specialist |
| paper, manuscript, regression | Academic research methodology specialist |
| project management, tracker, workflow | Operations and project management specialist |
| data, analysis, pipeline, code | Data science and reproducibility specialist |
| Default | Strategic planning and implementation specialist |

Override with `role:"..."` if provided.

Announce: "**Reviewing as:** Meticulous [role]."

### Step 3: Research Best Practices

Build web search queries:
- Query A: "[approach] best practices [year]"
- Query B: "[domain] common pitfalls"
- Query C (deep only): "[specific methodology] implementation guide"

| Depth | Searches |
|-------|----------|
| quick | 0 — skip |
| standard | 2 (A + B) |
| deep | 3-4 (all) |

Distill into 3-5 key principles relevant to this plan.

### Step 4: Structured Review

**CRITIC STANCE:** You are now the critic, not the planner. Do not rationalize. Your job is to find what's missing, what will break, and what's wishful thinking.

Review against 6 dimensions:

1. **Pre-mortem** — "It's 3 months later and this plan failed. Top 3 causes?"
2. **Completeness** — What's missing that a domain expert would expect?
3. **Feasibility** — Steps depending on unconfirmed resources or approvals?
4. **Best-practice alignment** — Compare to standards from Step 3
5. **Sequencing** — Hidden blockers? Would reordering reduce risk?
6. **Specificity** — Could someone unfamiliar execute each step?

**Classify findings:**
- **Red** — Critical. Will likely cause failure if unaddressed.
- **Yellow** — Important. Creates risk but plan can proceed.
- **Green** — Minor. Nice-to-have improvement.

### Step 5: Output

```
PLAN REVIEW — [Plan Title]

Reviewing as: Meticulous [role]
Plan source: [file / plan mode / conversation]
Depth: [quick / standard / deep]

BEST PRACTICES CONTEXT
[3-5 key principles from research]

STRENGTHS
1. [Label] — [Explanation]

WEAKNESSES & GAPS
[Red] [Label] — [Issue] → Fix: [Recommendation]
[Yellow] [Label] — [Issue] → Fix: [Recommendation]
[Green] [Label] — [Issue] → Fix: [Recommendation]

VERDICT
APPROVE — [Rationale]
  OR
REVISE — [Rationale]. See revised plan below.

REVISED PLAN (only if REVISE)
[Full revised plan with [CHANGED] and [NEW] markers]
```

### Step 6: Iteration

Ask: "Apply these revisions? Or provide feedback to refine further."
README.md

What This Does

Puts any plan through a structured expert review across 6 dimensions — pre-mortem analysis, completeness, feasibility, best-practice alignment, sequencing, and specificity. Catches blind spots, missing steps, and wishful thinking before they become real problems.


Quick Start

Step 1: Create Your Workspace

mkdir -p ~/Documents/Plans

Step 2: Download the Template

Click Download above, then:

mv ~/Downloads/CLAUDE.md ~/Documents/Plans/

Step 3: Run Claude Code

cd ~/Documents/Plans
claude

Then: "Review my plan" (if you just finished plan mode) or "Review the plan in project-plan.md"


The 6 Review Dimensions

Dimension What It Tests
Pre-mortem "It's 3 months later and this plan failed. What were the top 3 causes?"
Completeness What's missing that a domain expert would expect?
Feasibility Are there steps depending on unconfirmed resources or approvals?
Best-practice alignment How does this compare to industry standards?
Sequencing Are there hidden blockers? Would reordering reduce risk?
Specificity Could someone unfamiliar execute each step?

Finding Classification

Level Meaning Action
Red Critical — will likely cause failure if unaddressed Must fix before proceeding
Yellow Important — creates risk but plan can proceed Should fix soon
Green Minor — nice-to-have improvement Fix when convenient

Example Output

PLAN REVIEW — Q2 Product Launch

Reviewing as: Meticulous strategic planning specialist
Depth: standard

BEST PRACTICES CONTEXT
- Launch plans need rollback procedures
- Customer communication should precede internal deadlines
- Beta testing typically needs 2-4 weeks minimum

STRENGTHS
1. Clear ownership — every task has an assigned owner
2. Realistic timeline — builds in buffer for delays

WEAKNESSES & GAPS
[Red] No rollback plan — If launch fails, no procedure
  → Fix: Add rollback checklist with decision criteria
[Yellow] Beta period too short — 1 week may miss edge cases
  → Fix: Extend to 2 weeks or add staged rollout
[Green] Missing success metrics — No definition of "good launch"
  → Fix: Add KPIs for first 7/30/90 days

VERDICT
REVISE — Strong foundation but critical gap in rollback planning.

Options

Flag Syntax Default Purpose
File path file:path Auto-detect Explicit plan location
Expert role role:"..." Auto-detect Override reviewer persona
Focus area focus:dimension All Weight one dimension
Depth depth:quick/standard/deep standard Web research intensity
Dry run dryrun Off Show role + research plan only

Tips

  • Works on any plan — project plans, business strategies, product roadmaps, research proposals
  • Fresh-context review — Uses a subagent to avoid the bias of reviewing its own plan
  • Auto-detects domain — Assigns an appropriate expert role (AI engineering, grants, academic, operations, data science)
  • Pairs well with — Run after Claude Code's plan mode, or on any markdown plan file

Troubleshooting

"No plan found" Specify explicitly: "Review the plan in ~/Documents/my-plan.md"

Wrong expert role assigned Override: role:"marketing launch specialist"

Too many findings Use focus:feasibility to narrow to one dimension

$Related Playbooks