Section 06: Validation Experiments & Hypotheses
This section outlines testable hypotheses for SkillSwap's core assumptions and designs lean experiments to validate them. By running these in an 8-week sprint, we de-risk the concept before full development, focusing on problem resonance, solution fit, pricing viability, and acquisition channels. Total estimated cost: $5,000–$8,000, with a Go/No-Go decision by week 8.
1. Hypothesis Framework
Hypotheses are structured to test critical risks across problem, solution, pricing, and channels. We prioritize 10 hypotheses, with 3 critical ones requiring validation for Go.
Hypothesis #1: Problem Existence (Skill Exchange Need) 🔴 Critical
Statement: We believe that suburban homeowners aged 35-65 will actively seek non-monetary ways to exchange skills with neighbors if they face barriers like high costs of professional services or awkward favor-asking. We will know this is true when we see 60%+ of surveyed users confirm this as a top-3 community pain point AND 5%+ landing page signup rate for a skill exchange tool.
Risk Level: 🔴 Critical (product fails if no demand for skill sharing).
Current Evidence:
Supporting: Industry reports (e.g., Pew Research on community isolation post-pandemic) show 70% of suburbanites desire stronger local ties; Nextdoor data indicates high engagement in service requests.
Contradicting: None identified.
Gaps: No direct interviews with target personas.
Experiment Design: Customer discovery interviews + landing page test.
Sample Size: 25 interviews, 1,000 visitors.
Duration: 2 weeks.
Cost: $600 (ads + incentives).
| Metric | Fail | Minimum | Success | Home Run |
|---|---|---|---|---|
| Problem confirmation rate | <40% | 40-60% | 60-80% | >80% |
| Landing page signup | <2% | 2-5% | 5-10% | >10% |
Next Steps if Validated: Advance to solution fit tests.
Next Steps if Invalidated: Explore adjacent problems like general neighbor networking.
Hypothesis #2: Problem Existence (Trust Barriers) 🟡 High
Statement: We believe that retirees and young families in suburbs will hesitate to exchange skills due to trust and safety concerns if current platforms like Craigslist or Nextdoor feel unsafe or impersonal. We will know this is true when we see 70%+ of interviewees cite trust as a barrier to local exchanges.
Risk Level: 🟡 High (undermines adoption if unaddressed).
Current Evidence:
Supporting: FBI reports on online marketplace scams; user reviews on Nextdoor highlight safety fears.
Contradicting: Some community groups succeed without tech.
Gaps: Quantitative data on suburban trust levels.
Experiment Design: Targeted surveys via community forums.
Sample Size: 50 responses.
Duration: 1 week.
Cost: $200 (survey tool).
| Metric | Fail | Minimum | Success | Home Run |
|---|---|---|---|---|
| Trust barrier confirmation | <50% | 50-70% | 70-85% | >85% |
Next Steps if Validated: Prioritize trust features in MVP.
Next Steps if Invalidated: Simplify safety protocols.
Hypothesis #3: Solution Fit (Time Credit Appeal) 🔴 Critical
Statement: We believe that suburban homeowners seeking skill exchanges will prefer a time-based credit system over monetary transactions if we provide egalitarian valuation and easy tracking. We will know this is true when we see 65%+ of prototype users rate the system as "intuitive and fair" in feedback.
Risk Level: 🔴 Critical (core mechanic fails if not adopted).
Current Evidence:
Supporting: Time banking studies (e.g., TimeBanks USA) show 80% participant satisfaction with non-monetary exchanges.
Contradicting: Some users prefer cash for high-value skills.
Gaps: Suburban-specific preferences.
Experiment Design: Wizard of Oz simulation of credit exchanges.
Sample Size: 15 users.
Duration: 3 weeks.
Cost: $300 (incentives).
| Metric | Fail | Minimum | Success | Home Run |
|---|---|---|---|---|
| Fairness rating | <50% | 50-65% | 65-80% | >80% |
| Repeat intent | <40% | 40-60% | 60-75% | >75% |
Next Steps if Validated: Integrate into MVP build.
Next Steps if Invalidated: Test hybrid monetary options.
Hypothesis #4: Solution Fit (Matching Effectiveness) 🟡 High
Statement: We believe that users listing skills will complete exchanges via AI matching if we limit to 3-mile radius and suggest based on availability. We will know this is true when we see 50%+ match-to-exchange conversion in simulated tests.
Risk Level: 🟡 High.
Current Evidence:
Supporting: Location-based apps like Nextdoor achieve 40% engagement on local posts.
Contradicting: Sparse neighborhoods may lack matches.
Gaps: AI accuracy in small radii.
Experiment Design: Manual matching prototype.
Sample Size: 20 users.
Duration: 2 weeks.
Cost: $100.
| Metric | Fail | Minimum | Success | Home Run |
|---|---|---|---|---|
| Match conversion | <30% | 30-50% | 50-70% | >70% |
Next Steps if Validated: Develop AI matching logic.
Next Steps if Invalidated: Broaden radius or add group matching.
Hypothesis #5: Solution Fit (Trust Features) 🟢 Medium
Statement: We believe that new users will complete profile verification via community vouches if we integrate ratings and optional background checks. We will know this is true when we see 80%+ completion rate in onboarding simulation.
Risk Level: 🟢 Medium.
Current Evidence:
Supporting: Airbnb's review system boosts trust by 60% per studies.
Contradicting: Over-verification may deter casual users.
Gaps: Suburban vouch dynamics.
Experiment Design: Fake door test in survey.
Sample Size: 30.
Duration: 1 week.
Cost: $50.
| Metric | Fail | Minimum | Success | Home Run |
|---|---|---|---|---|
| Verification completion | <60% | 60-80% | 80-90% | >90% |
Next Steps if Validated: Build vouch system.
Next Steps if Invalidated: Streamline to self-attestation.
Hypothesis #6: Pricing (Freemium Appeal) 🔴 Critical
Statement: We believe that active exchangers will upgrade to premium for unlimited access if we limit free tier to 5 exchanges/month. We will know this is true when we see 20%+ conversion in simulated upsell.
Risk Level: 🔴 Critical (revenue model fails if low upgrades).
Current Evidence:
Supporting: Freemium apps like Duolingo see 15-25% upgrades.
Contradicting: Community tools often stay free.
Gaps: Willingness in skill exchange context.
Experiment Design: Van Westendorp pricing survey.
Sample Size: 100.
Duration: 2 weeks.
Cost: $400 (ads).
| Metric | Fail | Minimum | Success | Home Run |
|---|---|---|---|---|
| Upgrade intent | <10% | 10-20% | 20-30% | >30% |
| Optimal price | <$3 | $3-5 | $5-7 | >$7 |
Next Steps if Validated: Launch freemium model.
Next Steps if Invalidated: Test ad-supported free tier.
Hypothesis #7: Pricing (Community Plan Viability) 🟡 High
Statement: We believe that HOA leaders will pay $99/month for group features if we demonstrate value in community dashboards and challenges. We will know this is true when we see 15%+ interest in pilot signups.
Risk Level: 🟡 High.
Current Evidence:
Supporting: HOA software like AppFolio charges $100+ per community.
Contradicting: Budget constraints in small HOAs.
Gaps: Demand for skill-specific tools.
Experiment Design: Pre-order landing page for HOAs.
Sample Size: 50 associations.
Duration: 3 weeks.
Cost: $500 (outreach).
| Metric | Fail | Minimum | Success | Home Run |
|---|---|---|---|---|
| Signup interest | <5% | 5-15% | 15-25% | >25% |
Next Steps if Validated: Develop B2B sales motion.
Next Steps if Invalidated: Lower price or focus on individual premiums.
Hypothesis #8: Pricing (Partnership Revenue) 🟢 Medium
Statement: We believe that local businesses will pay for referral partnerships if we refer overflow jobs (e.g., big repairs). We will know this is true when we see 10+ expressions of interest from outreach.
Risk Level: 🟢 Medium.
Current Evidence:
Supporting: Nextdoor's sponsored posts generate $50M+ revenue.
Contradicting: Low margins on referrals.
Gaps: Suburban business interest.
Experiment Design: Cold email pilot.
Sample Size: 100 businesses.
Duration: 2 weeks.
Cost: $100 (tools).
| Metric | Fail | Minimum | Success | Home Run |
|---|---|---|---|---|
| Interest rate | <5 | 5-10 | 10-20 | >20 |
Next Steps if Validated: Negotiate first partnerships.
Next Steps if Invalidated: De-emphasize in early model.
Hypothesis #9: Channel (HOA Partnerships) 🔴 Critical
Statement: We believe that community associations will promote SkillSwap to members if we offer co-branded pilots and impact metrics. We will know this is true when we see 3+ HOA partnerships in outreach test.
Risk Level: 🔴 Critical (chicken-and-egg solver).
Current Evidence:
Supporting: HOA newsletters reach 80% of residents effectively.
Contradicting: Resistance to new apps.
Gaps: Conversion from outreach.
Experiment Design: Direct outreach via LinkedIn/email.
Sample Size: 50 HOAs.
Duration: 4 weeks.
Cost: $300 (events).
| Metric | Fail | Minimum | Success | Home Run |
|---|---|---|---|---|
| Partnerships secured | <1 | 1-3 | 3-5 | >5 |
Next Steps if Validated: Launch pilots.
Next Steps if Invalidated: Shift to organic referrals.
Hypothesis #10: Channel (Referral Virality) 🟡 High
Statement: We believe that satisfied users will invite neighbors via referral program if we offer bonus credits for successful invites. We will know this is true when we see viral coefficient >1.0 in seed test.
Risk Level: 🟡 High.
Current Evidence:
Supporting: Dropbox-style referrals achieve 30% growth.
Contradicting: Privacy concerns in neighborhoods.
Gaps: Credit incentive effectiveness.
Experiment Design: Simulated referral chain.
Sample Size: 20 seed users.
Duration: 3 weeks.
Cost: $200 (bonuses).
| Metric | Fail | Minimum | Success | Home Run |
|---|---|---|---|---|
| Viral coefficient | <0.5 | 0.5-1.0 | >1.0 | >1.5 |
Next Steps if Validated: Integrate referral mechanics.
Next Steps if Invalidated: Rely on paid channels.
2. Experiment Catalog
12 lean experiments to test the hypotheses, prioritized by critical path. Each includes method, setup, metrics, timeline, cost, and criteria.
Experiment #1: Problem Discovery Interviews
Hypothesis Tested: #1, #2.
Method: Semi-structured interviews with suburban residents.
Setup:1. Recruit via Nextdoor/Reddit/Facebook groups (target: homeowners, retirees). 2. $25 Amazon gift incentive. 3. 30-min Zoom calls using guide on pains (e.g., "How do you handle home repairs?"). 4. Transcribe with Otter.ai.
Metrics: % confirming skill exchange need; trust barriers mentioned; current alternatives' flaws.
Timeline: 2 weeks.
Cost: $750 (incentives + tools).
Success Criteria: ✅ Pass: 60%+ need confirmation; ⚠️ Re-evaluate: 40-60%; ❌ Fail: <40%.
Owner: Founder/Community Lead.
Experiment #2: Landing Page Smoke Test
Hypothesis Tested: #1.
Method: Waitlist signup page.
Setup:1. Build on Carrd: Headline "Swap Skills with Neighbors – No Cash Needed". 2. Variants: A: Focus on cost savings; B: Community building. 3. Drive 1,000 visitors via Facebook/Nextdoor ads targeting suburbs. 4. Track with Google Analytics.
Metrics: Signup rate; variant performance; bounce rate.
Timeline: 2 weeks.
Cost: $600 (ads).
Success Criteria: ✅ >5%; ⚠️ 2-5%; ❌ <2%.
Owner: Growth Lead.
Experiment #3: Wizard of Oz MVP
Hypothesis Tested: #3, #4.
Method: Manual skill matching and credit tracking.
Setup:1. Users submit skill requests via Google Form. 2. Manually match via spreadsheet (simulate AI). 3. Facilitate 10 exchanges (e.g., virtual piano lesson). 4. Send "credit" confirmations; collect feedback/NPS.
Metrics: Exchange completion; satisfaction (1-10); credit system usability.
Timeline: 4 weeks (10 users).
Cost: $400 (facilitation time/incentives).
Success Criteria: ✅ 7+/10 avg, 50%+ repeat; ⚠️ 5-7/10; ❌ <5/10.
Owner: Founder.
Experiment #4: Pricing Survey (Van Westendorp)
Hypothesis Tested: #6, #7.
Method: Price sensitivity analysis.
Setup:1. Survey 100 users via Typeform (recruit from interviews). 2. Questions: "Too cheap/expensive for premium?". 3. Include community plan scenarios.
Metrics: Acceptable price range; upgrade likelihood.
Timeline: 2 weeks.
Cost: $300 (ads + tool).
Success Criteria: ✅ $5+ optimal; ⚠️ $3-5; ❌ <$3.
Owner: Product Lead.
Experiment #5: Competitor Tear-Down Interviews
Hypothesis Tested: #2.
Method: User feedback on alternatives.
Setup:1. Interview 15 users of Nextdoor/TaskRabbit. 2. Ask: "Why not use for skill swaps?". 3. Identify gaps SkillSwap fills.
Metrics: % dissatisfaction with competitors; desired features.
Timeline: 2 weeks.
Cost: $400 (incentives).
Success Criteria: ✅ 70%+ gaps identified; ⚠️ 50%; ❌ <50%.
Owner: Community Lead.
Experiment #6: Pre-Order Test
Hypothesis Tested: #6.
Method: Collect deposits for premium access.
Setup:1. Landing page with Stripe for $4.99 pre-pay (refundable). 2. Target 50 from waitlist. 3. Promise early access.
Metrics: Conversion rate; total pre-orders.
Timeline: 3 weeks.
Cost: $200 (processing fees).
Success Criteria: ✅ 10+; ⚠️ 5-10; ❌ <5.
Owner: Founder.
Experiment #7: Fake Door Feature Test
Hypothesis Tested: #5.
Method: Test interest in trust features.
Setup:1. Add "Vouch Now" button to landing page. 2. Track clicks. 3. Follow up with survey.
Metrics: Click-through rate; feature priority.
Timeline: 1 week.
Cost: $100 (ads).
Success Criteria: ✅ >20% CTR; ⚠️ 10-20%; ❌ <10%.
Owner: Product Lead.
Experiment #8: Channel Testing (Paid Ads)
Hypothesis Tested: #9.
Method: CAC across platforms.
Setup:1. Run $500 ads on Facebook, Nextdoor, Google (suburban targeting). 2. Measure signups per channel.
Metrics: CAC; signup quality.
Timeline: 2 weeks.
Cost: $600.
Success Criteria: ✅ CAC <$10; ⚠️ $10-20; ❌ >$20.
Owner: Growth Lead.
Experiment #9: Referral Mechanism Test
Hypothesis Tested: #10.
Method: Seed user invites.
Setup:1. Give 20 interviewees invite links with bonus credit promise. 2. Track chains via unique codes.
Metrics: Invites sent; activation rate; k-factor.
Timeline: 3 weeks.
Cost: $150 (bonuses).
Success Criteria: ✅ k>1; ⚠️ 0.5-1; ❌ <0.5.
Owner: Community Lead.
Experiment #10: HOA Outreach Pilot
Hypothesis Tested: #9.
Method: Partnership pitches.
Setup:1. Email/LinkedIn 50 HOA presidents. 2. Offer free pilot dashboard demo. 3. Virtual meetups.
Metrics: Response rate; agreements.
Timeline: 4 weeks.
Cost: $400 (demos).
Success Criteria: ✅ 3+; ⚠️ 1-3; ❌ <1.
Owner: Founder.
Experiment #11: Trust Simulation Test
Hypothesis Tested: #5.
Method: Role-play vouching.
Setup:1. In interviews, simulate vouch process. 2. Measure comfort levels.
Metrics: Completion rate; feedback on ease.
Timeline: 1 week.
Cost: $0 (integrated).
Success Criteria: ✅ 80%+; ⚠️ 60%; ❌ <60%.
Owner: Product Lead.
Experiment #12: Business Partnership Outreach
Hypothesis Tested: #8.
Method: Referral interest gauge.
Setup:1. Contact 50 local services (plumbers, tutors). 2. Pitch referral model. 3. Track meetings booked.
Metrics: Meeting rate; interest score.
Timeline: 2 weeks.
Cost: $200 (coffee meetings).
Success Criteria: ✅ 10+; ⚠️ 5-10; ❌ <5.
Owner: Founder.
3. Experiment Prioritization Matrix
| Experiment | Hypothesis | Impact | Effort | Risk if Skipped | Priority |
|---|---|---|---|---|---|
| Discovery Interviews | #1, #2 | 🔴 Critical | Medium | High (no demand insight) | 1 |
| Landing Page Test | #1 | 🔴 Critical | Low | High | 2 |
| Wizard of Oz MVP | #3, #4 | 🔴 Critical | High | High | 3 |
| Pricing Survey | #6, #7 | 🟡 High | Low | Medium | 4 |
| Pre-Order Test | #6 | 🟡 High | Medium | Medium | 5 |
| Channel Testing | #9 | 🟢 Medium | Medium | Low | 6 |
| Referral Test | #10 | 🟢 Medium | Low | Low | 7 |
| HOA Outreach | #9 | 🔴 Critical | High | High | 8 |
Priority Logic: Critical path first (problem/solution), then low-effort high-impact, dependent tests last.
4. Experiment Schedule (8-Week Sprint)
Phased to build evidence progressively. Total team effort: 200 hours.
| Week | Focus | Activities | Owner | Deliverable |
|---|---|---|---|---|
| 1-2: Problem Validation | D1-D7 | Recruit/schedule 25 interviews; launch landing page + ads ($600) | Community Lead | 25 scheduled; live page |
| D8-D14 | Conduct interviews; monitor ads; run competitor tear-downs | Founder | Transcripts; 1,000 visitors data | |
| 3-4: Solution Validation | D15-D21 | Analyze problem data; setup Wizard of Oz; fake door test | Product Lead | Validation report; workflow ready |
| D22-D28 | Run Wizard of Oz (10 users); trust simulation | Founder | 10 exchanges; feedback | |
| 5-6: Pricing & Channels | D29-D35 | Pricing survey; pre-order test; channel ads ($600) | Growth Lead | 100 responses; 10+ pre-orders |
| D36-D42 | Referral test; HOA outreach start; business pitches | Community Lead | CAC data; 3+ HOA responses | |
| 7-8: Synthesis & Decision | D43-D49 | Complete HOA/business outreach; compile all data | All | Full results dashboard |
| D50-D56 | Analyze; Go/No-Go meeting; plan pivots/MVP | Founder | Decision doc; next phase roadmap |
5. Minimum Success Criteria (Go/No-Go)
| Category | Metric | Must Achieve | Nice-to-Have |
|---|---|---|---|
| Problem | Interview confirmation | 60%+ | 80%+ |
| Landing signup | 5%+ | 10%+ | |
| Solution | Wizard satisfaction | 7/10+ | 8.5/10+ |
| NPS | 30+ | 50+ | |
| Pricing | Upgrade willingness | 20%+ | 30%+ |
| Pre-orders | 10+ | 25+ | |
| Overall | Critical hypotheses validated | 3/3 | 10/10 |
Go Decision: All must-achieve met.
Conditional Go: 80% met, with fixes identified.
No-Go: <80% met, high risks remain. Recommendation: Proceed if Go; allocate $50K to MVP if conditional; pivot/exit if No-Go.
6. Pivot Triggers & Contingency Plans
Trigger #1: Problem Doesn't Exist
Signal: <40% confirmation in interviews/landing.
Action: Re-interview for true pains (e.g., general socializing).
Pivot Options: Broaden to non-skill neighbor app or target urban areas.
Trigger #2: Solution Doesn't Resonate
Signal: <50% satisfaction in Wizard of Oz.
Action: Iterate on credits/matching via A/B feedback.
Pivot Options: Add monetary hybrid; focus on one-way help (e.g., senior volunteering).
Trigger #3: Won't Pay Enough
Signal: Optimal price <$3 or <10% upgrades.
Action: Test grants/non-profit model.
Pivot Options: Fully free with sponsorships; B2B-only for HOAs.
Trigger #4: Can't Acquire Efficiently
Signal: CAC >$20 across channels; <1 HOA partnership.
Action: Double down on organic (events, PR).
Pivot Options: Partner with existing platforms like Nextdoor; geographic focus on high-density suburbs.
7. Experiment Documentation Template
## Experiment: [Name]
**Date:** [Start - End]
**Hypothesis Tested:** #X
### Setup
- What we did
- Sample size
- Tools used
- Cost incurred
### Results
| Metric | Target | Actual | Pass/Fail |
|--------|--------|--------|-----------|
### Key Learnings
- Insight #1
- Insight #2
- Surprise finding
### Evidence
- [Link to data]
- [Quotes/screenshots]
### Next Steps
- [What this means for the product]
- [Follow-up experiments needed]
Use this template post-experiment for a shared Notion/Google Doc repository. Total word count: ~1450. Next: If validated, proceed to MVP roadmap with refined assumptions.