Agent skill

validation-frameworks

Problem and solution validation methodologies, assumption testing, and MVP validation experiments

Stars 163
Forks 31

Install this agent skill to your Project

npx add-skill https://github.com/majiayu000/claude-skill-registry/tree/main/skills/testing/validation-frameworks-slgoodrich-agents-45736bf9

SKILL.md

Validation Frameworks

Frameworks for validating problems, solutions, and product assumptions before committing to full development.

When to Use This Skill

Auto-loaded by agents:

  • research-ops - For problem/solution validation and assumption testing

Use when you need:

  • Validating product ideas before building
  • Testing assumptions and hypotheses
  • De-risking product decisions
  • Running MVP experiments
  • Validating problem-solution fit
  • Testing willingness to pay
  • Evaluating technical feasibility

Core Principle: Reduce Risk Through Learning

Building the wrong thing is expensive. Validation reduces risk by answering critical questions before major investment:

  • Problem validation: Is this a real problem worth solving?
  • Solution validation: Does our solution actually solve the problem?
  • Market validation: Will people pay for this?
  • Usability validation: Can people use it?
  • Technical validation: Can we build it?

The Validation Spectrum

Validation isn't binary (validated vs. not validated). It's a spectrum of confidence:

Wild Guess → Hypothesis → Validated Hypothesis → High Confidence → Proven

Early stage: Cheap, fast tests (low confidence gain) Later stage: More expensive tests (high confidence gain)

Example progression:

  1. Assumption: "Busy parents struggle to plan healthy meals"
  2. Interview 5 parents → Some validation (small sample)
  3. Survey 100 parents → More validation (larger sample)
  4. Prototype test with 20 parents → Strong validation (behavior observed)
  5. Launch MVP, track engagement → Very strong validation (real usage)
  6. Measure retention after 3 months → Proven (sustained behavior)

Problem Validation vs. Solution Validation

Problem Validation

Question: "Is this a problem worth solving?"

Goal: Confirm that:

  • The problem exists
  • It's painful enough that people want it solved
  • It's common enough to matter
  • Current solutions are inadequate

Methods:

  • Customer interviews (discovery-focused)
  • Ethnographic observation
  • Surveys about pain points
  • Data analysis (support tickets, analytics)
  • Jobs-to-be-Done interviews

Red flags that stop here:

  • No one cares about this problem
  • Existing solutions work fine
  • Problem only affects tiny niche
  • Pain point is mild annoyance, not real pain

Output: Problem statement + evidence it's worth solving

See assets/problem-validation-canvas.md for a ready-to-use framework.


Solution Validation

Question: "Does our solution solve the problem?"

Goal: Confirm that:

  • Our solution addresses the problem
  • Users understand it
  • Users would use it
  • It's better than alternatives
  • Value proposition resonates

Methods:

  • Prototype testing
  • Landing page tests
  • Concierge MVP (manual solution)
  • Wizard of Oz (fake backend)
  • Pre-sales or waitlist signups

Red flags:

  • Users don't understand the solution
  • They prefer their current workaround
  • Would use it but not pay for it
  • Solves wrong part of the problem

Output: Validated solution concept + evidence of demand

See assets/solution-validation-checklist.md for validation steps.


The Assumption-Validation Cycle

1. Identify Assumptions

Every product idea rests on assumptions. Make them explicit.

Types of assumptions:

Desirability: Will people want this?

  • "Users want to track their habits"
  • "They'll pay $10/month for this"
  • "They'll invite their friends"

Feasibility: Can we build this?

  • "We can process data in under 1 second"
  • "We can integrate with their existing tools"
  • "Our team can build this in 3 months"

Viability: Should we build this?

  • "Customer acquisition cost will be under $50"
  • "Retention will be above 40% after 30 days"
  • "We can reach 10k users in 12 months"

Best practice: Write assumptions as testable hypotheses

  • Not: "Users need this feature"
  • Yes: "At least 60% of users will use this feature weekly"

2. Prioritize Assumptions to Test

Not all assumptions are equally important. Prioritize by:

Risk: How uncertain are we? (Higher risk = higher priority) Impact: What happens if we're wrong? (Higher impact = higher priority)

Prioritization matrix:

Risk/Impact High Impact Low Impact
High Risk Test first Test second
Low Risk Test second Maybe skip

Riskiest assumptions (test these first):

  • Leap-of-faith assumptions the entire business depends on
  • Things you've never done before
  • Assumptions about user behavior with no data
  • Technical feasibility of core functionality

Lower-risk assumptions (test later or assume):

  • Based on prior experience
  • Easy to change if wrong
  • Industry best practices
  • Minor features

3. Design Validation Experiments

For each assumption, design the cheapest test that could prove it wrong.

Experiment design principles:

1. Falsifiable: Can produce evidence that assumption is wrong 2. Specific: Clear success/failure criteria defined upfront 3. Minimal: Smallest effort needed to learn 4. Fast: Results in days/weeks, not months 5. Ethical: Don't mislead or harm users

The Experiment Canvas: See assets/validation-experiment-template.md


4. Run the Experiment

Before starting:

  • Define success criteria ("At least 40% will click")
  • Set sample size ("Test with 50 users")
  • Decide timeframe ("Run for 1 week")
  • Identify what success/failure would mean for product

During:

  • Track metrics rigorously
  • Document unexpected learnings
  • Don't change experiment mid-flight

After:

  • Analyze results honestly (avoid confirmation bias)
  • Document what you learned
  • Decide: Pivot, persevere, or iterate

Validation Methods by Fidelity

Low-Fidelity (Hours to Days)

1. Customer Interviews

  • Cost: Very low (just time)
  • Validates: Problem existence, pain level, current solutions
  • Limitations: What people say ≠ what they do
  • Best for: Early problem validation

2. Surveys

  • Cost: Low
  • Validates: Problem prevalence, feature preferences
  • Limitations: Response bias, can't probe deeply
  • Best for: Quantifying what you learned qualitatively

3. Landing Page Test

  • Cost: Low (1-2 days to build)
  • Validates: Interest in solution, value proposition clarity
  • Measure: Email signups, clicks to "Get Started"
  • Best for: Demand validation before building

4. Paper Prototypes

  • Cost: Very low (sketch on paper/whiteboard)
  • Validates: Concept understanding, workflow feasibility
  • Limitations: Low realism
  • Best for: Very early solution concepts

Medium-Fidelity (Days to Weeks)

5. Clickable Prototypes

  • Cost: Medium (design tool, 2-5 days)
  • Validates: User flow, interaction patterns, comprehension
  • Tools: Figma, Sketch, Adobe XD
  • Best for: Usability validation pre-development

6. Concierge MVP

  • Cost: Medium (your time delivering manually)
  • Validates: Value of outcome, willingness to use/pay
  • Example: Manually curate recommendations before building algorithm
  • Best for: Validating value before automation

7. Wizard of Oz MVP

  • Cost: Medium (build facade, manual backend)
  • Validates: User behavior, feature usage, workflows
  • Example: "AI" feature that's actually humans behind the scenes
  • Best for: Testing expensive-to-build features

High-Fidelity (Weeks to Months)

8. Functional Prototype

  • Cost: High (weeks of development)
  • Validates: Technical feasibility, actual usage patterns
  • Limitations: Expensive if you're wrong
  • Best for: After other validation, final pre-launch check

9. Private Beta

  • Cost: High (full build + support)
  • Validates: Real-world usage, retention, bugs
  • Best for: Pre-launch final validation with early adopters

10. Public MVP

  • Cost: Very high (full product)
  • Validates: Product-market fit, business model viability
  • Best for: After all other validation passed

Validation Metrics and Success Criteria

Quantitative Validation Metrics

Interest/Demand:

  • Landing page conversion rate
  • Waitlist signup rate
  • Pre-order conversion
  • Prototype test completion rate

Usage:

  • Feature adoption rate
  • Task completion rate
  • Time on task
  • Frequency of use

Engagement:

  • DAU/MAU (daily/monthly active users)
  • Session length
  • Return rate
  • Feature usage depth

Value Realization:

  • Aha moment rate (reached key action)
  • Time to value
  • Retention curves
  • Referral rate

Business:

  • Conversion to paid
  • Lifetime value (LTV)
  • Customer acquisition cost (CAC)
  • Churn rate

Qualitative Validation Signals

Strong positive signals:

  • Users ask when they can buy it
  • Users describe it to others excitedly
  • Users find use cases you didn't imagine
  • Users return without prompting
  • Users complain when you take it away

Weak positive signals:

  • "That's interesting"
  • "I might use that"
  • Polite feedback with no follow-up
  • Theoretical interest but no commitment

Negative signals:

  • Confusion about value proposition
  • Preference for existing solution
  • Feature requests that miss the point
  • Can't articulate who would use it
  • Ghosting after initial interest

Setting Success Criteria

Before running experiment, define what success looks like.

Framework: Set three thresholds

  1. Strong success: Clear green light, proceed confidently
  2. Moderate success: Promising but needs iteration
  3. Failure: Pivot or abandon

Example: Landing page test

  • Strong success: > 30% email signup rate
  • Moderate success: 15-30% signup rate
  • Failure: < 15% signup rate

Example: Prototype test

  • Strong success: 8/10 users complete task, would use weekly
  • Moderate success: 5-7/10 complete, would use monthly
  • Failure: < 5/10 complete or no usage intent

Important: Decide criteria before seeing results to avoid bias.


Teresa Torres Continuous Discovery Validation

Opportunity Solution Trees

Map opportunities (user needs) to solutions to validate:

Outcome
└── Opportunity 1
    ├── Solution A
    ├── Solution B
    └── Solution C
└── Opportunity 2
    └── Solution D

Validate each level:

  1. Outcome: Is this the right goal?
  2. Opportunities: Are these real user needs?
  3. Solutions: Will this solution address the opportunity?

Assumption Testing

For each solution, map assumptions:

Desirability assumptions: Will users want this? Usability assumptions: Can users use this? Feasibility assumptions: Can we build this? Viability assumptions: Should we build this?

Then test riskiest assumptions first with smallest possible experiments.

Weekly Touchpoints

Continuous discovery = continuous validation:

  • Weekly customer interviews (problem + solution validation)
  • Weekly prototype tests with 2-3 users
  • Weekly assumption tests (small experiments)

Goal: Continuous evidence flow, not one-time validation.

See references/lean-startup-validation.md and references/assumption-testing-methods.md for detailed methodologies.


Common Validation Anti-Patterns

1. Fake Validation

What it looks like:

  • Asking friends and family (they'll lie to be nice)
  • Leading questions ("Wouldn't you love...?")
  • Testing with employees
  • Cherry-picking positive feedback

Fix: Talk to real users, ask open-ended questions, seek disconfirming evidence.


2. Analysis Paralysis

What it looks like:

  • Endless research without decisions
  • Testing everything before building anything
  • Demanding statistical significance with 3 data points

Fix: Accept uncertainty, set decision deadlines, bias toward action.


3. Confirmation Bias

What it looks like:

  • Only hearing what confirms existing beliefs
  • Dismissing negative feedback as "they don't get it"
  • Stopping research when you hear what you wanted

Fix: Actively seek disconfirming evidence, set falsifiable criteria upfront.


4. Vanity Validation

What it looks like:

  • "I got 500 email signups!" (but 0 conversions)
  • "People loved the demo!" (but won't use it)
  • "We got great feedback!" (but all feature requests, no usage)

Fix: Focus on behavior over opinions, retention over acquisition.


5. Building Instead of Validating

What it looks like:

  • "Let's build it and see if anyone uses it"
  • "It'll only take 2 weeks" (takes 2 months)
  • Full build before any user contact

Fix: Always do cheapest possible test first, build only after validation.


Validation Workflow: End-to-End Example

Idea: Mobile app for tracking baby sleep patterns

Phase 1: Problem Validation

Assumption: "New parents struggle to understand their baby's sleep patterns"

Experiment 1: Customer Interviews

  • Interview 10 new parents
  • Ask about sleep tracking current methods, pain points
  • Success criteria: 7/10 express frustration with current methods

Result: 9/10 keep notes on paper, all frustrated by inconsistency

Conclusion: Problem validated, proceed.


Phase 2: Solution Validation (Concept)

Assumption: "Parents would use an app that auto-tracks sleep from smart monitor"

Experiment 2: Landing Page

  • Create page describing solution
  • Run ads targeting new parents
  • Success criteria: 20% email signup rate

Result: 8% signup rate

Conclusion: Interest is lower than hoped. Talk to non-signups to understand why.

Learning: They don't have smart monitors, manual entry would be fine.


Phase 3: Solution Validation (Refined)

Assumption: "Parents will manually log sleep times if it's fast"

Experiment 3: Concierge MVP

  • Ask 5 parents to text sleep times
  • Manually create charts/insights
  • Send back daily summaries
  • Success criteria: 4/5 use it for 2 weeks

Result: All 5 use it for 2 weeks, ask to keep using

Conclusion: Strong validation, build simple MVP.


Phase 4: Build & Validate MVP

Build: Simple app with manual time entry + basic charts

Experiment 4: Private Beta

  • 30 parents use MVP for 4 weeks
  • Success criteria: 60% retention after 2 weeks, 40% after 4 weeks

Result: 70% 2-week retention, 50% 4-week retention

Conclusion: Exceeded criteria, prepare for launch.


This workflow: problem → concept → refined solution → MVP took 8 weeks and cost minimal development before validation.


Validation Checklist by Stage

Idea Stage

  • Problem validated through customer interviews
  • Current solutions identified and evaluated
  • Target user segment defined and accessible
  • Pain level assessed (nice-to-have vs. must-have)

Concept Stage

  • Solution concept tested with users
  • Value proposition resonates
  • Demand signal measured (signups, interest)
  • Key assumptions identified and prioritized

Pre-Build Stage

  • Prototype tested with target users
  • Core workflow validated
  • Pricing validated (willingness to pay)
  • Technical feasibility confirmed

MVP Stage

  • Beta users recruited
  • Usage patterns observed
  • Retention measured
  • Unit economics validated

Ready-to-Use Resources

In assets/:

  • problem-validation-canvas.md: Framework for validating problems before solutions
  • solution-validation-checklist.md: Step-by-step checklist for solution validation
  • validation-experiment-template.md: Design experiments to test assumptions

In references/:

  • lean-startup-validation.md: Build-Measure-Learn cycle, MVP types, pivot decisions
  • assumption-testing-methods.md: Comprehensive assumption testing techniques

Further Reading

Books:

  • "The Lean Startup" by Eric Ries
  • "The Mom Test" by Rob Fitzpatrick
  • "Continuous Discovery Habits" by Teresa Torres
  • "Testing Business Ideas" by Bland & Osterwalder

Key Concepts:

  • Minimum Viable Product (MVP)
  • Build-Measure-Learn
  • Validated learning
  • Pivot or persevere
  • Assumption mapping
  • Evidence-based product development

Remember: Validation isn't about proving you're right. It's about learning what's true. Seek evidence that you're wrong - that's the fastest path to building something valuable.

Didn't find tool you were looking for?

Be as detailed as possible for better results