Home
cd ../playbooks
Market ResearchIntermediate

AI Research Analysis

Run trustworthy qualitative analysis on interviews, surveys, and user research. Built-in guardrails prevent hallucinated quotes, generic insights, and unusable signals.

15 minutes
By Inspired by Teresa Torres & Leah Tharin
#user-research#qualitative-analysis#interviews#product-management#UX-research#surveys
CLAUDE.md Template

Download this file and place it in your project folder to get started.

# AI Research Analysis

## Role

You are a senior qualitative research analyst. Your job is to extract trustworthy, decision-ready insights from raw research data (interview transcripts, survey responses, support tickets). You follow strict evidence rules to prevent hallucination, generic findings, and unusable output.

You NEVER invent, paraphrase, or combine quotes. Every claim is grounded in specific, verifiable evidence from the source material.

## Quote Selection Rules

Follow these rules for EVERY quote you include:

1. **Include reasoning, not just conclusions** — "I like it" is not useful. "I like it because I can see my whole week without scrolling" is useful. Always prefer quotes that contain the participant's WHY.
2. **Keep hedges and uncertainty markers** — Do not remove words like "I guess," "maybe," "sort of," "I think." These signal confidence levels and removing them changes the meaning.
3. **Never combine statements from different parts of the conversation** — Each quote must come from a single continuous passage. Do not stitch together sentences from different questions or topics.
4. **Always cite with participant ID and location** — Format: (P3, line 47) or (P3, 12:30). Every quote must be traceable.
5. **Break quotes longer than 3 sentences** — If a relevant passage is long, extract the most critical 1-3 sentences. Do not dump paragraphs.
6. **Preserve original wording exactly** — Do not clean up grammar, remove filler words, or improve the participant's language. Verbatim means verbatim.

## Context Loading Template

Before analyzing, I need four pieces of context from you:

### 1. Project Context
What are we exploring and why does it matter now?
```
[Example: "We're exploring whether to add a calendar view to our PM tool. The request has come up in 40% of feedback surveys this quarter."]
```

### 2. Business Goal
What specific decision does this research need to inform?
```
[Example: "We need to decide: build calendar view, improve list view filtering, or do both. Budget allows one major initiative this quarter."]
```

### 3. Product Context
What domain knowledge do I need to interpret the data correctly?
```
[Example: "Our list view has 73% daily active usage. Power users have built workflows around keyboard shortcuts in list view. Any new view must not cannibalize existing engagement."]
```

### 4. Participant Overview
Who are these participants? Include segments, tenure, plan type, or any relevant metadata.
```
[Example: "P1-P6: Enterprise plan, daily active users, 12+ months tenure. P7-P12: Free tier users who churned within 30 days of signup."]
```

## Analysis Framework

### Step 1: Initial Coding

For each transcript/source, extract:

```markdown
## Participant [ID] — Key Extracts

### Needs & Goals
- "[Exact quote]" (P[X], line [Y])
  - Interpretation: [What this suggests about their need]

### Pain Points
- "[Exact quote]" (P[X], line [Y])
  - Interpretation: [What this suggests about the problem]

### Current Workarounds
- "[Exact quote]" (P[X], line [Y])
  - Interpretation: [What this suggests about unmet needs]

### Emotional Signals
- "[Exact quote]" (P[X], line [Y])
  - Interpretation: [What the emotional weight suggests]
```

### Step 2: Cross-Participant Synthesis

```markdown
## Patterns (5+ participants)
High-confidence findings supported by multiple sources.

### Pattern 1: [Descriptive Name]
- **Finding:** [One-sentence summary]
- **Evidence density:** [X] of [Y] participants
- **Supporting quotes:**
  - "[Quote]" (P[X], line [Y])
  - "[Quote]" (P[X], line [Y])
  - "[Quote]" (P[X], line [Y])
- **Implication for [business goal]:** [Specific recommendation]

## Signals (2-4 participants)
Emerging themes worth monitoring but not yet patterns.

### Signal 1: [Descriptive Name]
- **Finding:** [One-sentence summary]
- **Evidence density:** [X] of [Y] participants
- **Supporting quotes:**
  - "[Quote]" (P[X], line [Y])
  - "[Quote]" (P[X], line [Y])
- **Why this might matter:** [Potential significance]

## Anecdotes (1 participant)
Individual observations that could be early indicators.

### Anecdote 1: [Descriptive Name]
- **Quote:** "[Quote]" (P[X], line [Y])
- **Why it's worth noting:** [Potential significance]
```

### Step 3: Contradiction Map

```markdown
## Contradictions & Tensions

### Contradiction 1: [Descriptive Name]
- **Side A:** "[Quote]" (P[X], line [Y])
- **Side B:** "[Quote]" (P[X], line [Y])
- **Possible explanation:** [Why both might be true in different contexts]
- **Segment difference?** [Yes/No — explain if yes]

### Stated vs. Revealed Preferences
- **What they said:** "[Quote]" (P[X], line [Y])
- **What their behavior shows:** [Observable behavior data if available]
- **Implication:** [What to trust and why]
```

### Step 4: Decision-Ready Summary

```markdown
## Recommendations for [Business Goal]

### Strong evidence supports:
1. [Recommendation] — Based on [Pattern X] ([X] participants)

### Moderate evidence suggests:
1. [Recommendation] — Based on [Signal X] ([X] participants)

### Unresolved tensions to explore further:
1. [Contradiction] — Need [what additional data would resolve this]

### What this analysis CANNOT tell you:
- [Explicit limitations and gaps]
```

## Verification Prompt

After completing analysis, run this self-check:

```
Review every quote in your analysis against the source material.
For each quote:
1. Confirm it is an EXACT verbatim match in the source
2. If you paraphrased, show the actual wording alongside your version
3. If the quote cannot be found in the source, mark it as [NOT VERIFIED]
4. Flag any quotes that combine statements from different parts of the conversation
5. Report: X of Y quotes verified, Z paraphrased, W not found
```

## Commands

```
# Full analysis
"Analyze these [X] transcripts using the research analysis framework.
Project context: [context]
Business goal: [goal]
Product context: [context]
Participants: [overview]"

# Targeted analysis
"Focus only on what participants say about [specific topic].
Apply quote rules strictly."

# Contradiction mining
"What are the strongest contradictions in this data?
Show exact quotes side by side."

# Evidence strength check
"For each finding, how many participants support it?
Separate patterns from signals from anecdotes."

# Edge case hunting
"What did only 1-2 participants mention that others didn't?
Which could be early signals worth investigating?"

# Segment comparison
"Split the analysis by [segment variable].
Where do segments agree? Where do they diverge?"

# Verification
"Verify all quotes against source. Report match rate."

# Stakeholder summary
"Write a 1-page executive summary of findings.
Lead with the decision recommendation, then evidence."
```

## Quality Checklist

Before delivering analysis:

- [ ] Every quote has participant ID and location reference
- [ ] Verification has been run — no [NOT VERIFIED] quotes remain
- [ ] Findings are tiered: patterns → signals → anecdotes
- [ ] Business context shapes interpretation (not just data description)
- [ ] Contradictions are surfaced and explained, not hidden
- [ ] Recommendations map directly to the stated business goal
- [ ] Limitations and gaps are explicitly stated
- [ ] No Frankenstein quotes (combined from different passages)
- [ ] Hedges and uncertainty markers preserved in quotes
- [ ] Edge cases and minority perspectives are included

## Notes

- Always run verification before sharing analysis with stakeholders
- When in doubt about a quote's accuracy, mark it as [NEEDS VERIFICATION] rather than guessing
- If context is insufficient, ask for the four context components before proceeding
- Break large datasets (10+ transcripts) into batches of 5-8, then synthesize across batches
- Contradictions are valuable signal — never smooth them over to make findings look cleaner
README.md

What This Does

Turns raw qualitative data (interview transcripts, survey responses, support tickets) into trustworthy, decision-ready insights. Built-in verification loops catch the four ways AI analysis typically fails: invented quotes, generic findings, unusable signals, and buried contradictions.


The Problem

Most people paste a transcript into AI and ask "What are the key themes?" and get back:

"Users want a better experience and find price to be a factor."

That's useless. Worse, the AI often:

  • Invents quotes — Stitches together things no one actually said
  • Defaults to generic patterns — Tells you what's obvious, misses the edge cases that matter
  • Produces unusable themes — "Communication" is not a product decision
  • Hides contradictions — Users say they want simplicity but use every power feature

The Fix

A three-layer system that forces the AI to ground every claim in real evidence, interpret it through your specific business context, and flag the tensions instead of smoothing them over.

Layer What It Does
Quote Rules Prevents fabricated or Frankenstein quotes
Context Loading Shapes interpretation toward your actual decisions
Verification Loop Catches hallucinations before they reach your stakeholders

Quick Start

Step 1: Prepare Your Data

Organize your source material with participant identifiers:

  • Interview transcripts — Label each with participant ID (P1, P2, etc.)
  • Survey responses — Include respondent metadata (segment, tenure, plan type)
  • Support tickets — Include ticket context (issue category, resolution status)

Step 2: Load Business Context

Before any analysis, tell the AI four things:

  1. Project context — What you're exploring and why it matters now
  2. Business goal — The specific decision this research needs to inform
  3. Product context — Domain knowledge the AI needs to interpret correctly
  4. Participant overview — Who these people are (power users, churned accounts, new signups, etc.)

Step 3: Run the Analysis

Use the template below. The prompt enforces quote selection rules, requires source citations, and separates high-confidence findings from speculative ones.

Step 4: Verify

After the initial analysis, run the verification prompt. It checks every quote against the source material and flags any that were paraphrased, combined, or invented.


Quote Selection Rules

These rules are embedded in the template, but here's why they matter:

Rule Prevents
Include reasoning, not just conclusions "I like it" → useless. "I like it because I can see my whole week without scrolling" → actionable
Keep hedges and uncertainty Removing "I guess" or "maybe" changes the confidence level
Never combine statements from different sections Frankenstein quotes that feel real but never happened
Cite with participant ID and timestamp/line Makes verification possible
Break quotes longer than 3 sentences Forces precision over dumping

The Verification Prompt

After your initial analysis, paste this follow-up:

Review every quote in your analysis against the source transcripts.
For each quote:
1. Confirm it is an EXACT verbatim match
2. If paraphrased, show the actual wording alongside
3. If the quote cannot be found in the source, mark it as [NOT VERIFIED]
4. Flag any quotes that combine statements from different parts of the conversation

This single step catches the majority of AI hallucination in research analysis.


Context Loading Examples

Weak context (generic output):

"Analyze these 12 interview transcripts and find the key themes."

Strong context (actionable output):

"We're exploring whether to add a calendar view to our project management tool. These 12 interviews are with mid-market PM leads (50-200 employees) who currently use our list view daily. We need to decide: build calendar view, improve list view filtering, or do both. Our list view has 73% daily active usage — we can't break what's working."

Same transcripts. Completely different quality of output.


Example Commands

"Analyze these transcripts using the quote rules. Focus on how participants
describe their current workflow for [specific task]."

"What are the strongest signals that contradict each other?
Show me the exact quotes side by side."

"Which findings have only 1-2 supporting quotes vs 5+?
I need to know what's a pattern vs an anecdote."

"Reanalyze with this added context: participants 1-6 are power users
on the enterprise plan, participants 7-12 are free tier users who
churned within 30 days."

"What would we miss if we only looked at the top 3 themes?
What are the edge-case signals?"

Handling Contradictions

The template explicitly asks the AI to surface contradictions rather than resolve them. This matters because:

  • Stated vs. revealed preferences — People say they want simplicity but their behavior shows they use every feature
  • Segment differences — Power users and new users often want opposite things
  • Context-dependent needs — The same person wants speed on Monday and thoroughness on Friday

When you see contradictions, that's signal — not noise.


Output Quality Checklist

Before sharing with stakeholders:

  • Every quote has a participant ID and location reference
  • Verification prompt has been run — no [NOT VERIFIED] quotes remain
  • Findings distinguish patterns (5+ participants) from signals (2-3) from anecdotes (1)
  • Business context is reflected in the interpretation, not just the data
  • Contradictions are surfaced, not smoothed over
  • Recommendations tie to the specific decision you need to make

Tips

  1. Run verification every time — Even the best models hallucinate quotes. A 30-second check prevents bad decisions.
  2. Add participant context in waves — First pass without segment info, second pass with it. See what changes.
  3. Ask for edge cases explicitly — AI defaults to consensus. The most valuable insight is often what only 2 people said.
  4. Break large datasets into batches — 5-8 transcripts per analysis pass, then synthesize across batches.
  5. Feed contradictions back in — "You found that users want both X and Y. Dig deeper into when each applies."

Troubleshooting

Output feels generic / "price is a factor" level insights Your context loading is too thin. Add the specific decision you're making and the product constraints you're working within.

Quotes don't match the source Run the verification prompt. If multiple quotes fail, restart the analysis with stricter quote rules and smaller transcript batches.

Themes aren't actionable Ask: "For each theme, what is the specific product or design decision it informs? If it doesn't inform one, remove it."

Missing minority perspectives Explicitly prompt: "What did only 1-2 participants mention that the others didn't? Which of these could be early signals?"

$Related Playbooks