Home
cd ../playbooks
Product ManagementIntermediate

Proof of Life (PoL) Probe

Define and document a lightweight, disposable validation artifact designed to surface harsh truths before expensive development — the artifact template that pairs with pol-probe-advisor.

15 minutes
By communitySource
#validation#prototyping#discovery#pol-probe#lean#experiments

Your prototype has three weeks of dev time in it, a real database, and a staging environment — and you haven't deleted anything yet. That's not a probe, that's a half-built product. A PoL probe has a disposal date, a single hypothesis, and a stinging success criterion that makes you wince if you fail. Reconnaissance, not MVP.

Who it's for: PMs documenting validation artifacts, UX leads running discovery sprints, founders avoiding prototype theater, teams committing to disposable experiments, engineers allocating spike time

Example

"Document a PoL probe for testing 3-field onboarding vs. 8-field" → Hypothesis + risk + probe type (Task-Focused) + 10 waitlist users + success criteria (8+ complete <2min / <6 or >5min fail) + tools (Maze + Loom) + 5-day timeline + disposal plan

CLAUDE.md Template

New here? 3-minute setup guide → | Already set up? Copy the template below.

# Proof of Life (PoL) Probe

Define and document a **Proof of Life probe** — a lightweight, disposable validation artifact designed to surface harsh truths before expensive development. Reconnaissance missions, not MVPs. Meant to be **deleted, not scaled**.

Prevents prototype theater. Forces method to match actual learning goal.

## What Is a PoL Probe?

A deliberate, disposable validation experiment answering one specific question as cheaply and quickly as possible. Not a product, not an MVP, not a pilot — a targeted truth-seeking mission.

**Origin:** Coined by Dean Peters (Productside), building on Marty Cagan's 2014 prototype flavors and Jeff Patton: *"The most expensive way to test your idea is to build production-quality software."*

## 5 Essential Characteristics

| Characteristic | Meaning | Why |
|----------------|---------|-----|
| **Lightweight** | Hours/days, not weeks | Expensive probes resist deletion |
| **Disposable** | Planned for deletion, not scaling | Prevents sunk-cost fallacy, scope creep |
| **Narrow Scope** | One hypothesis or risk | Broad experiments yield ambiguous results |
| **Brutally Honest** | Harsh truths, not vanity metrics | Polite data is useless data |
| **Tiny & Focused** | Reconnaissance, never MVPs | Small surface = faster learning |

**Anti-Pattern:** If it feels too polished to delete, it's prototype theater.

## PoL Probe vs. MVP

| Dimension | PoL Probe | MVP |
|-----------|-----------|-----|
| Purpose | De-risk via narrow hypothesis | Justify ideas / defend roadmap |
| Scope | Single question, single risk | Smallest shippable increment |
| Lifespan | Hours-days, deleted | Weeks-months, iterated |
| Audience | Internal team + narrow sample | Real customers in production |
| Fidelity | Just enough illusion | Production-quality |
| Outcome | Learn what *doesn't* work | Learn what *does* work (and ship) |

**Key:** PoL probes are **pre-MVP reconnaissance**. Run probes to decide IF you should build an MVP.

## The 5 Prototype Flavors

| Type | Core Question | Timeline | Tools |
|------|---------------|----------|-------|
| **Feasibility Check** | "Can we build this?" | 1-2 days | GenAI chains, API tests, data sweeps, spike code |
| **Task-Focused Test** | "Can users complete without friction?" | 2-5 days | Optimal Workshop, UsabilityHub, Maze |
| **Narrative Prototype** | "Earn stakeholder buy-in?" | 1-3 days | Loom, Sora/Synthesia, slideware, storyboards |
| **Synthetic Data Simulation** | "Model without production risk?" | 2-4 days | Synthea, DataStax LangFlow |
| **Vibe-Coded Probe** | "Survive real user contact?" | 2-3 days | ChatGPT Canvas + Replit + Airtable |

**Golden Rule:** *"Use the cheapest prototype that tells the harshest truth. If it doesn't sting, it's probably theater."*

## When to Use

**Use:** Specific falsifiable hypothesis, particular risk blocks next decision, need harsh truth in days not weeks, production software would be premature, can articulate what failure looks like.

**Don't:** Impressing execs (theater), confirmation bias, can't articulate disposal plan, broad learning goal ("will customers like this?"), avoiding hard decision.

## The Template

```markdown
# PoL Probe: [Descriptive Name]

## Hypothesis
[One-sentence statement of what you believe]
Example: "If we reduce onboarding to 3 fields, completion will exceed 80%."

## Risk Being Eliminated
[Specific risk/unknown]
Example: "We don't know if users abandon signup due to form length."

## Prototype Type
- [ ] Feasibility Check
- [ ] Task-Focused Test
- [ ] Narrative Prototype
- [ ] Synthetic Data Simulation
- [x] Vibe-Coded PoL Probe

## Target Users / Audience
[Who interacts with this probe?]
Example: "10 users from early access waitlist, non-technical SMB owners."

## Success Criteria (Harsh Truth)
- **Pass:** 8+ users complete signup under 2 minutes
- **Fail:** <6 users complete, or average time exceeds 5 minutes
- **Learn:** Identify specific drop-off fields

## Tools / Stack
Example: "ChatGPT Canvas for form UI, Airtable for data, Loom for post-session interviews."

## Timeline
- Build: 2 days
- Test: 1 day (10 user sessions)
- Analyze: 1 day
- Disposal: Day 5 (delete code, keep learnings doc)

## Disposal Plan
Example: "After sessions, archive recordings, delete Frankensoft code, document learnings in Notion."

## Owner
[Who is accountable for running and disposing]

## Status
- [ ] Hypothesis defined
- [ ] Probe built
- [ ] Users recruited
- [ ] Testing complete
- [ ] Learnings documented
- [ ] Probe disposed
```

## Quality Checklist

Before launching, verify:

- [ ] **Lightweight:** Built in 1-3 days?
- [ ] **Disposable:** Disposal date committed?
- [ ] **Narrow:** Tests ONE hypothesis?
- [ ] **Brutally Honest:** Will data hurt if wrong?
- [ ] **Tiny:** Smaller than MVP?
- [ ] **Falsifiable:** Can describe failure?
- [ ] **Clear Owner:** One person accountable?

If any "no," revise or reconsider.

## Example (Mini)

```markdown
Hypothesis: Users can distinguish "archive" vs. "delete"
Probe Type: Task-Focused Test
Pass: 80%+ correct interpretation
Fail: <60% or critical confusion
Disposal: Archive test results after 10 sessions
```

## Common Pitfalls

1. Running "will users like this?" — too broad; use falsifiable hypothesis
2. Treating probe as proto-MVP, refusing disposal
3. Vanity metrics that avoid uncomfortable truth
4. Skipping pre-defined failure threshold
5. Choosing tools first, hypothesis second

## References

- `pol-probe-advisor` — Decision framework for choosing probe type
- `discovery-process` — Use probes in validation phase
- `problem-statement` — Define problem first
- `epic-hypothesis` — Frame hypothesis
- Jeff Patton — *User Story Mapping*
- Marty Cagan — *Inspired* (2014 prototype flavors)
- Dean Peters — [*Vibe First, Validate Fast, Verify Fit*](https://deanpeters.substack.com/p/vibe-first-validate-fast-verify-fit)
README.md

What This Does

Template for documenting a Proof of Life probe with 5 essential characteristics (lightweight, disposable, narrow scope, brutally honest, tiny & focused), explicit disposal plan, and falsifiable success/fail criteria.

Pairs with pol-probe-advisor (decision tree for which probe type) and discovery-process (where probes live in the cycle).


Quick Start

mkdir -p ~/Documents/PolProbe
mv ~/Downloads/CLAUDE.md ~/Documents/PolProbe/
cd ~/Documents/PolProbe
claude

Provide hypothesis, risk, timeline, and chosen probe type. Claude produces the fully filled template with success criteria, tools, and disposal plan.


The 5 Essential Characteristics

Characteristic Meaning
Lightweight Hours/days, not weeks
Disposable Planned for deletion, not scaling
Narrow Scope One hypothesis, one risk
Brutally Honest Harsh truths, not vanity metrics
Tiny & Focused Reconnaissance, never MVP

If your probe feels too polished to delete, it's prototype theater.


PoL Probe vs. MVP

  • PoL Probe: Hours-days, disposable, tests IF you should build
  • MVP: Weeks-months, iterated, ships what you build

Probes are pre-MVP reconnaissance.


Tips & Best Practices

  • Set disposal date before building. Without one, the probe becomes a product.
  • Write success/fail/learn criteria before building. "We'll know it when we see it" = no truth.
  • One probe, one hypothesis. Bundle 3 questions = ambiguous results.
  • Choose method AFTER hypothesis. Tools-first biases toward what you can build, not what you need to learn.
  • Celebrate the jank. Frankensoft is the point — a clean probe defeats its purpose.

Common Pitfalls

  • Running broad "will users like this?" tests (not falsifiable)
  • Treating the probe like a proto-MVP and refusing to delete it
  • Vanity metrics that avoid uncomfortable truth
  • No pre-defined failure threshold before testing starts
  • Choosing tools before defining the hypothesis

$Related Playbooks