diff --git a/product-team/product-manager-toolkit/SKILL.md b/product-team/product-manager-toolkit/SKILL.md index f0d605b..b5d1181 100644 --- a/product-team/product-manager-toolkit/SKILL.md +++ b/product-team/product-manager-toolkit/SKILL.md @@ -7,11 +7,32 @@ description: Comprehensive toolkit for product managers including RICE prioritiz Essential tools and frameworks for modern product management, from discovery to delivery. +--- + +## Table of Contents + +- [Quick Start](#quick-start) +- [Core Workflows](#core-workflows) + - [Feature Prioritization](#feature-prioritization-process) + - [Customer Discovery](#customer-discovery-process) + - [PRD Development](#prd-development-process) +- [Tools Reference](#tools-reference) + - [RICE Prioritizer](#rice-prioritizer) + - [Customer Interview Analyzer](#customer-interview-analyzer) +- [Input/Output Examples](#inputoutput-examples) +- [Integration Points](#integration-points) +- [Common Pitfalls](#common-pitfalls-to-avoid) + +--- + ## Quick Start ### For Feature Prioritization ```bash -python scripts/rice_prioritizer.py sample # Create sample CSV +# Create sample data file +python scripts/rice_prioritizer.py sample + +# Run prioritization with team capacity python scripts/rice_prioritizer.py sample_features.csv --capacity 15 ``` @@ -22,318 +43,443 @@ python scripts/customer_interview_analyzer.py interview_transcript.txt ### For PRD Creation 1. Choose template from `references/prd_templates.md` -2. Fill in sections based on discovery work -3. Review with stakeholders -4. Version control in your PM tool +2. Fill sections based on discovery work +3. Review with engineering for feasibility +4. Version control in project management tool + +--- ## Core Workflows ### Feature Prioritization Process -1. **Gather Feature Requests** - - Customer feedback - - Sales requests - - Technical debt - - Strategic initiatives +``` +Gather → Score → Analyze → Plan → Validate → Execute +``` -2. **Score with RICE** - ```bash - # Create CSV with: name,reach,impact,confidence,effort - python scripts/rice_prioritizer.py features.csv - ``` - - **Reach**: Users affected per quarter - - **Impact**: massive/high/medium/low/minimal - - **Confidence**: high/medium/low - - **Effort**: xl/l/m/s/xs (person-months) +#### Step 1: Gather Feature Requests +- Customer feedback (support tickets, interviews) +- Sales requests (CRM pipeline blockers) +- Technical debt (engineering input) +- Strategic initiatives (leadership goals) -3. **Analyze Portfolio** - - Review quick wins vs big bets - - Check effort distribution - - Validate against strategy +#### Step 2: Score with RICE +```bash +# Input: CSV with features +python scripts/rice_prioritizer.py features.csv --capacity 20 +``` -4. **Generate Roadmap** - - Quarterly capacity planning - - Dependency mapping - - Stakeholder alignment +See `references/frameworks.md` for RICE formula and scoring guidelines. + +#### Step 3: Analyze Portfolio +Review the tool output for: +- Quick wins vs big bets distribution +- Effort concentration (avoid all XL projects) +- Strategic alignment gaps + +#### Step 4: Generate Roadmap +- Quarterly capacity allocation +- Dependency identification +- Stakeholder communication plan + +#### Step 5: Validate Results +**Before finalizing the roadmap:** +- [ ] Compare top priorities against strategic goals +- [ ] Run sensitivity analysis (what if estimates are wrong by 2x?) +- [ ] Review with key stakeholders for blind spots +- [ ] Check for missing dependencies between features +- [ ] Validate effort estimates with engineering + +#### Step 6: Execute and Iterate +- Share roadmap with team +- Track actual vs estimated effort +- Revisit priorities quarterly +- Update RICE inputs based on learnings + +--- ### Customer Discovery Process -1. **Conduct Interviews** - - Use semi-structured format - - Focus on problems, not solutions - - Record with permission +``` +Plan → Recruit → Interview → Analyze → Synthesize → Validate +``` -2. **Analyze Insights** - ```bash - python scripts/customer_interview_analyzer.py transcript.txt - ``` - Extracts: - - Pain points with severity - - Feature requests with priority - - Jobs to be done - - Sentiment analysis - - Key themes and quotes +#### Step 1: Plan Research +- Define research questions +- Identify target segments +- Create interview script (see `references/frameworks.md`) -3. **Synthesize Findings** - - Group similar pain points - - Identify patterns across interviews - - Map to opportunity areas +#### Step 2: Recruit Participants +- 5-8 interviews per segment +- Mix of power users and churned users +- Incentivize appropriately -4. **Validate Solutions** - - Create solution hypotheses - - Test with prototypes - - Measure actual vs expected behavior +#### Step 3: Conduct Interviews +- Use semi-structured format +- Focus on problems, not solutions +- Record with permission +- Take minimal notes during interview + +#### Step 4: Analyze Insights +```bash +python scripts/customer_interview_analyzer.py transcript.txt +``` + +Extracts: +- Pain points with severity +- Feature requests with priority +- Jobs to be done patterns +- Sentiment and key themes +- Notable quotes + +#### Step 5: Synthesize Findings +- Group similar pain points across interviews +- Identify patterns (3+ mentions = pattern) +- Map to opportunity areas using Opportunity Solution Tree +- Prioritize opportunities by frequency and severity + +#### Step 6: Validate Solutions +**Before building:** +- [ ] Create solution hypotheses (see `references/frameworks.md`) +- [ ] Test with low-fidelity prototypes +- [ ] Measure actual behavior vs stated preference +- [ ] Iterate based on feedback +- [ ] Document learnings for future research + +--- ### PRD Development Process -1. **Choose Template** - - **Standard PRD**: Complex features (6-8 weeks) - - **One-Page PRD**: Simple features (2-4 weeks) - - **Feature Brief**: Exploration phase (1 week) - - **Agile Epic**: Sprint-based delivery - -2. **Structure Content** - - Problem → Solution → Success Metrics - - Always include out-of-scope - - Clear acceptance criteria - -3. **Collaborate** - - Engineering for feasibility - - Design for experience - - Sales for market validation - - Support for operational impact - -## Key Scripts - -### rice_prioritizer.py -Advanced RICE framework implementation with portfolio analysis. - -**Features**: -- RICE score calculation -- Portfolio balance analysis (quick wins vs big bets) -- Quarterly roadmap generation -- Team capacity planning -- Multiple output formats (text/json/csv) - -**Usage Examples**: -```bash -# Basic prioritization -python scripts/rice_prioritizer.py features.csv - -# With custom team capacity (person-months per quarter) -python scripts/rice_prioritizer.py features.csv --capacity 20 - -# Output as JSON for integration -python scripts/rice_prioritizer.py features.csv --output json +``` +Scope → Draft → Review → Refine → Approve → Track ``` -### customer_interview_analyzer.py +#### Step 1: Choose Template +Select from `references/prd_templates.md`: + +| Template | Use Case | Timeline | +|----------|----------|----------| +| Standard PRD | Complex features, cross-team | 6-8 weeks | +| One-Page PRD | Simple features, single team | 2-4 weeks | +| Feature Brief | Exploration phase | 1 week | +| Agile Epic | Sprint-based delivery | Ongoing | + +#### Step 2: Draft Content +- Lead with problem statement +- Define success metrics upfront +- Explicitly state out-of-scope items +- Include wireframes or mockups + +#### Step 3: Review Cycle +- Engineering: feasibility and effort +- Design: user experience gaps +- Sales: market validation +- Support: operational impact + +#### Step 4: Refine Based on Feedback +- Address technical constraints +- Adjust scope to fit timeline +- Document trade-off decisions + +#### Step 5: Approval and Kickoff +- Stakeholder sign-off +- Sprint planning integration +- Communication to broader team + +#### Step 6: Track Execution +**After launch:** +- [ ] Compare actual metrics vs targets +- [ ] Conduct user feedback sessions +- [ ] Document what worked and what didn't +- [ ] Update estimation accuracy data +- [ ] Share learnings with team + +--- + +## Tools Reference + +### RICE Prioritizer + +Advanced RICE framework implementation with portfolio analysis. + +**Features:** +- RICE score calculation with configurable weights +- Portfolio balance analysis (quick wins vs big bets) +- Quarterly roadmap generation based on capacity +- Multiple output formats (text, JSON, CSV) + +**CSV Input Format:** +```csv +name,reach,impact,confidence,effort,description +User Dashboard Redesign,5000,high,high,l,Complete redesign +Mobile Push Notifications,10000,massive,medium,m,Add push support +Dark Mode,8000,medium,high,s,Dark theme option +``` + +**Commands:** +```bash +# Create sample data +python scripts/rice_prioritizer.py sample + +# Run with default capacity (10 person-months) +python scripts/rice_prioritizer.py features.csv + +# Custom capacity +python scripts/rice_prioritizer.py features.csv --capacity 20 + +# JSON output for integration +python scripts/rice_prioritizer.py features.csv --output json + +# CSV output for spreadsheets +python scripts/rice_prioritizer.py features.csv --output csv +``` + +--- + +### Customer Interview Analyzer + NLP-based interview analysis for extracting actionable insights. -**Capabilities**: +**Capabilities:** - Pain point extraction with severity assessment - Feature request identification and classification - Jobs-to-be-done pattern recognition -- Sentiment analysis -- Theme extraction -- Competitor mentions -- Key quotes identification +- Sentiment analysis per section +- Theme and quote extraction +- Competitor mention detection -**Usage Examples**: +**Commands:** ```bash -# Analyze single interview +# Analyze interview transcript python scripts/customer_interview_analyzer.py interview.txt -# Output as JSON for aggregation +# JSON output for aggregation python scripts/customer_interview_analyzer.py interview.txt json ``` -## Reference Documents +--- -### prd_templates.md -Multiple PRD formats for different contexts: +## Input/Output Examples -1. **Standard PRD Template** - - Comprehensive 11-section format - - Best for major features - - Includes technical specs +### RICE Prioritizer Example -2. **One-Page PRD** - - Concise format for quick alignment - - Focus on problem/solution/metrics - - Good for smaller features - -3. **Agile Epic Template** - - Sprint-based delivery - - User story mapping - - Acceptance criteria focus - -4. **Feature Brief** - - Lightweight exploration - - Hypothesis-driven - - Pre-PRD phase - -## Prioritization Frameworks - -### RICE Framework -``` -Score = (Reach × Impact × Confidence) / Effort - -Reach: # of users/quarter -Impact: - - Massive = 3x - - High = 2x - - Medium = 1x - - Low = 0.5x - - Minimal = 0.25x -Confidence: - - High = 100% - - Medium = 80% - - Low = 50% -Effort: Person-months +**Input (features.csv):** +```csv +name,reach,impact,confidence,effort +Onboarding Flow,20000,massive,high,s +Search Improvements,15000,high,high,m +Social Login,12000,high,medium,m +Push Notifications,10000,massive,medium,m +Dark Mode,8000,medium,high,s ``` -### Value vs Effort Matrix -``` - Low Effort High Effort - -High QUICK WINS BIG BETS -Value [Prioritize] [Strategic] - -Low FILL-INS TIME SINKS -Value [Maybe] [Avoid] +**Command:** +```bash +python scripts/rice_prioritizer.py features.csv --capacity 15 ``` -### MoSCoW Method -- **Must Have**: Critical for launch -- **Should Have**: Important but not critical -- **Could Have**: Nice to have -- **Won't Have**: Out of scope - -## Discovery Frameworks - -### Customer Interview Guide +**Output:** ``` -1. Context Questions (5 min) - - Role and responsibilities - - Current workflow - - Tools used +============================================================ +RICE PRIORITIZATION RESULTS +============================================================ -2. Problem Exploration (15 min) - - Pain points - - Frequency and impact - - Current workarounds +📊 TOP PRIORITIZED FEATURES -3. Solution Validation (10 min) - - Reaction to concepts - - Value perception - - Willingness to pay +1. Onboarding Flow + RICE Score: 16000.0 + Reach: 20000 | Impact: massive | Confidence: high | Effort: s -4. Wrap-up (5 min) - - Other thoughts - - Referrals - - Follow-up permission +2. Search Improvements + RICE Score: 4800.0 + Reach: 15000 | Impact: high | Confidence: high | Effort: m + +3. Social Login + RICE Score: 3072.0 + Reach: 12000 | Impact: high | Confidence: medium | Effort: m + +4. Push Notifications + RICE Score: 3840.0 + Reach: 10000 | Impact: massive | Confidence: medium | Effort: m + +5. Dark Mode + RICE Score: 2133.33 + Reach: 8000 | Impact: medium | Confidence: high | Effort: s + +📈 PORTFOLIO ANALYSIS + +Total Features: 5 +Total Effort: 19 person-months +Total Reach: 65,000 users +Average RICE Score: 5969.07 + +🎯 Quick Wins: 2 features + • Onboarding Flow (RICE: 16000.0) + • Dark Mode (RICE: 2133.33) + +🚀 Big Bets: 0 features + +📅 SUGGESTED ROADMAP + +Q1 - Capacity: 11/15 person-months + • Onboarding Flow (RICE: 16000.0) + • Search Improvements (RICE: 4800.0) + • Dark Mode (RICE: 2133.33) + +Q2 - Capacity: 10/15 person-months + • Push Notifications (RICE: 3840.0) + • Social Login (RICE: 3072.0) ``` -### Hypothesis Template +--- + +### Customer Interview Analyzer Example + +**Input (interview.txt):** ``` -We believe that [building this feature] -For [these users] -Will [achieve this outcome] -We'll know we're right when [metric] +Customer: Jane, Enterprise PM at TechCorp +Date: 2024-01-15 + +Interviewer: What's the hardest part of your current workflow? + +Jane: The biggest frustration is the lack of real-time collaboration. +When I'm working on a PRD, I have to constantly ping my team on Slack +to get updates. It's really frustrating to wait for responses, +especially when we're on a tight deadline. + +I've tried using Google Docs for collaboration, but it doesn't +integrate with our roadmap tools. I'd pay extra for something that +just worked seamlessly. + +Interviewer: How often does this happen? + +Jane: Literally every day. I probably waste 30 minutes just on +back-and-forth messages. It's my biggest pain point right now. ``` -### Opportunity Solution Tree -``` -Outcome -├── Opportunity 1 -│ ├── Solution A -│ └── Solution B -└── Opportunity 2 - ├── Solution C - └── Solution D +**Command:** +```bash +python scripts/customer_interview_analyzer.py interview.txt ``` -## Metrics & Analytics - -### North Star Metric Framework -1. **Identify Core Value**: What's the #1 value to users? -2. **Make it Measurable**: Quantifiable and trackable -3. **Ensure It's Actionable**: Teams can influence it -4. **Check Leading Indicator**: Predicts business success - -### Funnel Analysis Template +**Output:** ``` -Acquisition → Activation → Retention → Revenue → Referral +============================================================ +CUSTOMER INTERVIEW ANALYSIS +============================================================ -Key Metrics: -- Conversion rate at each step -- Drop-off points -- Time between steps -- Cohort variations +📋 INTERVIEW METADATA +Segments found: 1 +Lines analyzed: 15 + +😟 PAIN POINTS (3 found) + +1. [HIGH] Lack of real-time collaboration + "I have to constantly ping my team on Slack to get updates" + +2. [MEDIUM] Tool integration gaps + "Google Docs...doesn't integrate with our roadmap tools" + +3. [HIGH] Time wasted on communication + "waste 30 minutes just on back-and-forth messages" + +💡 FEATURE REQUESTS (2 found) + +1. Real-time collaboration - Priority: High +2. Seamless tool integration - Priority: Medium + +🎯 JOBS TO BE DONE + +When working on PRDs with tight deadlines +I want real-time visibility into team updates +So I can avoid wasted time on status checks + +📊 SENTIMENT ANALYSIS + +Overall: Negative (pain-focused interview) +Key emotions: Frustration, Time pressure + +💬 KEY QUOTES + +• "It's really frustrating to wait for responses" +• "I'd pay extra for something that just worked seamlessly" +• "It's my biggest pain point right now" + +🏷️ THEMES + +- Collaboration friction +- Tool fragmentation +- Time efficiency ``` -### Feature Success Metrics -- **Adoption**: % of users using feature -- **Frequency**: Usage per user per time period -- **Depth**: % of feature capability used -- **Retention**: Continued usage over time -- **Satisfaction**: NPS/CSAT for feature - -## Best Practices - -### Writing Great PRDs -1. Start with the problem, not solution -2. Include clear success metrics upfront -3. Explicitly state what's out of scope -4. Use visuals (wireframes, flows) -5. Keep technical details in appendix -6. Version control changes - -### Effective Prioritization -1. Mix quick wins with strategic bets -2. Consider opportunity cost -3. Account for dependencies -4. Buffer for unexpected work (20%) -5. Revisit quarterly -6. Communicate decisions clearly - -### Customer Discovery Tips -1. Ask "why" 5 times -2. Focus on past behavior, not future intentions -3. Avoid leading questions -4. Interview in their environment -5. Look for emotional reactions -6. Validate with data - -### Stakeholder Management -1. Identify RACI for decisions -2. Regular async updates -3. Demo over documentation -4. Address concerns early -5. Celebrate wins publicly -6. Learn from failures openly - -## Common Pitfalls to Avoid - -1. **Solution-First Thinking**: Jumping to features before understanding problems -2. **Analysis Paralysis**: Over-researching without shipping -3. **Feature Factory**: Shipping features without measuring impact -4. **Ignoring Technical Debt**: Not allocating time for platform health -5. **Stakeholder Surprise**: Not communicating early and often -6. **Metric Theater**: Optimizing vanity metrics over real value +--- ## Integration Points -This toolkit integrates with: -- **Analytics**: Amplitude, Mixpanel, Google Analytics -- **Roadmapping**: ProductBoard, Aha!, Roadmunk -- **Design**: Figma, Sketch, Miro -- **Development**: Jira, Linear, GitHub -- **Research**: Dovetail, UserVoice, Pendo -- **Communication**: Slack, Notion, Confluence +Compatible tools and platforms: -## Quick Commands Cheat Sheet +| Category | Platforms | +|----------|-----------| +| **Analytics** | Amplitude, Mixpanel, Google Analytics | +| **Roadmapping** | ProductBoard, Aha!, Roadmunk, Productplan | +| **Design** | Figma, Sketch, Miro | +| **Development** | Jira, Linear, GitHub, Asana | +| **Research** | Dovetail, UserVoice, Pendo, Maze | +| **Communication** | Slack, Notion, Confluence | + +**JSON export enables integration with most tools:** +```bash +# Export for Jira import +python scripts/rice_prioritizer.py features.csv --output json > priorities.json + +# Export for dashboard +python scripts/customer_interview_analyzer.py interview.txt json > insights.json +``` + +--- + +## Common Pitfalls to Avoid + +| Pitfall | Description | Prevention | +|---------|-------------|------------| +| **Solution-First** | Jumping to features before understanding problems | Start every PRD with problem statement | +| **Analysis Paralysis** | Over-researching without shipping | Set time-boxes for research phases | +| **Feature Factory** | Shipping features without measuring impact | Define success metrics before building | +| **Ignoring Tech Debt** | Not allocating time for platform health | Reserve 20% capacity for maintenance | +| **Stakeholder Surprise** | Not communicating early and often | Weekly async updates, monthly demos | +| **Metric Theater** | Optimizing vanity metrics over real value | Tie metrics to user value delivered | + +--- + +## Best Practices + +**Writing Great PRDs:** +- Start with the problem, not the solution +- Include clear success metrics upfront +- Explicitly state what's out of scope +- Use visuals (wireframes, flows, diagrams) +- Keep technical details in appendix +- Version control all changes + +**Effective Prioritization:** +- Mix quick wins with strategic bets +- Consider opportunity cost of delays +- Account for dependencies between features +- Buffer 20% for unexpected work +- Revisit priorities quarterly +- Communicate decisions with context + +**Customer Discovery:** +- Ask "why" five times to find root cause +- Focus on past behavior, not future intentions +- Avoid leading questions ("Wouldn't you love...") +- Interview in the user's natural environment +- Watch for emotional reactions (pain = opportunity) +- Validate qualitative with quantitative data + +--- + +## Quick Reference ```bash # Prioritization @@ -342,10 +488,17 @@ python scripts/rice_prioritizer.py features.csv --capacity 15 # Interview Analysis python scripts/customer_interview_analyzer.py interview.txt -# Create sample data +# Generate sample data python scripts/rice_prioritizer.py sample -# JSON outputs for integration +# JSON outputs python scripts/rice_prioritizer.py features.csv --output json python scripts/customer_interview_analyzer.py interview.txt json ``` + +--- + +## Reference Documents + +- `references/prd_templates.md` - PRD templates for different contexts +- `references/frameworks.md` - Detailed framework documentation (RICE, MoSCoW, Kano, JTBD, etc.) diff --git a/product-team/product-manager-toolkit/references/frameworks.md b/product-team/product-manager-toolkit/references/frameworks.md new file mode 100644 index 0000000..24c250d --- /dev/null +++ b/product-team/product-manager-toolkit/references/frameworks.md @@ -0,0 +1,559 @@ +# Product Management Frameworks + +Comprehensive reference for prioritization, discovery, and measurement frameworks. + +--- + +## Table of Contents + +- [Prioritization Frameworks](#prioritization-frameworks) + - [RICE Framework](#rice-framework) + - [Value vs Effort Matrix](#value-vs-effort-matrix) + - [MoSCoW Method](#moscow-method) + - [ICE Scoring](#ice-scoring) + - [Kano Model](#kano-model) +- [Discovery Frameworks](#discovery-frameworks) + - [Customer Interview Guide](#customer-interview-guide) + - [Hypothesis Template](#hypothesis-template) + - [Opportunity Solution Tree](#opportunity-solution-tree) + - [Jobs to Be Done](#jobs-to-be-done) +- [Metrics Frameworks](#metrics-frameworks) + - [North Star Metric](#north-star-metric-framework) + - [HEART Framework](#heart-framework) + - [Funnel Analysis](#funnel-analysis-template) + - [Feature Success Metrics](#feature-success-metrics) +- [Strategic Frameworks](#strategic-frameworks) + - [Product Vision Template](#product-vision-template) + - [Competitive Analysis](#competitive-analysis-framework) + - [Go-to-Market Checklist](#go-to-market-checklist) + +--- + +## Prioritization Frameworks + +### RICE Framework + +**Formula:** +``` +RICE Score = (Reach × Impact × Confidence) / Effort +``` + +**Components:** + +| Component | Description | Values | +|-----------|-------------|--------| +| **Reach** | Users affected per quarter | Numeric count (e.g., 5000) | +| **Impact** | Effect on each user | massive=3x, high=2x, medium=1x, low=0.5x, minimal=0.25x | +| **Confidence** | Certainty in estimates | high=100%, medium=80%, low=50% | +| **Effort** | Person-months required | xl=13, l=8, m=5, s=3, xs=1 | + +**Example Calculation:** +``` +Feature: Mobile Push Notifications +Reach: 10,000 users +Impact: massive (3x) +Confidence: medium (80%) +Effort: medium (5 person-months) + +RICE = (10,000 × 3 × 0.8) / 5 = 4,800 +``` + +**Interpretation Guidelines:** +- **1000+**: High priority - strong candidates for next quarter +- **500-999**: Medium priority - consider for roadmap +- **100-499**: Low priority - keep in backlog +- **<100**: Deprioritize - requires new data to reconsider + +**When to Use RICE:** +- Quarterly roadmap planning +- Comparing features across different product areas +- Communicating priorities to stakeholders +- Resolving prioritization debates with data + +**RICE Limitations:** +- Requires reasonable estimates (garbage in, garbage out) +- Doesn't account for dependencies +- May undervalue platform investments +- Reach estimates can be gaming-prone + +--- + +### Value vs Effort Matrix + +``` + Low Effort High Effort + +--------------+------------------+ + High Value | QUICK WINS | BIG BETS | + | [Do First] | [Strategic] | + +--------------+------------------+ + Low Value | FILL-INS | TIME SINKS | + | [Maybe] | [Avoid] | + +--------------+------------------+ +``` + +**Quadrant Definitions:** + +| Quadrant | Characteristics | Action | +|----------|-----------------|--------| +| **Quick Wins** | High impact, low effort | Prioritize immediately | +| **Big Bets** | High impact, high effort | Plan strategically, validate ROI | +| **Fill-Ins** | Low impact, low effort | Use to fill sprint gaps | +| **Time Sinks** | Low impact, high effort | Avoid unless required | + +**Portfolio Balance:** +- Ideal mix: 40% Quick Wins, 30% Big Bets, 20% Fill-Ins, 10% Buffer +- Review balance quarterly +- Adjust based on team morale and strategic goals + +--- + +### MoSCoW Method + +| Category | Definition | Sprint Allocation | +|----------|------------|-------------------| +| **Must Have** | Critical for launch; product fails without it | 60% of capacity | +| **Should Have** | Important but workarounds exist | 20% of capacity | +| **Could Have** | Desirable enhancements | 10% of capacity | +| **Won't Have** | Explicitly out of scope (this release) | 0% - documented | + +**Decision Criteria for "Must Have":** +- Regulatory/legal requirement +- Core user job cannot be completed without it +- Explicitly promised to customers +- Security or data integrity requirement + +**Common Mistakes:** +- Everything becomes "Must Have" (scope creep) +- Not documenting "Won't Have" items +- Treating "Should Have" as optional (they're important) +- Forgetting to revisit for next release + +--- + +### ICE Scoring + +**Formula:** +``` +ICE Score = (Impact + Confidence + Ease) / 3 +``` + +| Component | Scale | Description | +|-----------|-------|-------------| +| **Impact** | 1-10 | Expected effect on key metric | +| **Confidence** | 1-10 | How sure are you about impact? | +| **Ease** | 1-10 | How easy to implement? | + +**When to Use ICE vs RICE:** +- ICE: Early-stage exploration, quick estimates +- RICE: Quarterly planning, cross-team prioritization + +--- + +### Kano Model + +Categories of feature satisfaction: + +| Type | Absent | Present | Priority | +|------|--------|---------|----------| +| **Basic (Must-Be)** | Dissatisfied | Neutral | High - table stakes | +| **Performance (Linear)** | Neutral | Satisfied proportionally | Medium - differentiation | +| **Excitement (Delighter)** | Neutral | Very satisfied | Strategic - competitive edge | +| **Indifferent** | Neutral | Neutral | Low - skip unless cheap | +| **Reverse** | Satisfied | Dissatisfied | Avoid - remove if exists | + +**Feature Classification Questions:** +1. How would you feel if the product HAS this feature? +2. How would you feel if the product DOES NOT have this feature? + +--- + +## Discovery Frameworks + +### Customer Interview Guide + +**Structure (35 minutes total):** + +``` +1. CONTEXT QUESTIONS (5 min) + └── Build rapport, understand role + +2. PROBLEM EXPLORATION (15 min) + └── Dig into pain points + +3. SOLUTION VALIDATION (10 min) + └── Test concepts if applicable + +4. WRAP-UP (5 min) + └── Referrals, follow-up +``` + +**Detailed Script:** + +#### Phase 1: Context (5 min) +``` +"Thanks for taking the time. Before we dive in..." + +- What's your role and how long have you been in it? +- Walk me through a typical day/week. +- What tools do you use for [relevant task]? +``` + +#### Phase 2: Problem Exploration (15 min) +``` +"I'd love to understand the challenges you face with [area]..." + +- What's the hardest part about [task]? +- Can you tell me about the last time you struggled with this? +- What did you do? What happened? +- How often does this happen? +- What does it cost you (time, money, frustration)? +- What have you tried to solve it? +- Why didn't those solutions work? +``` + +#### Phase 3: Solution Validation (10 min) +``` +"Based on what you've shared, I'd like to get your reaction to an idea..." + +[Show prototype/concept - keep it rough to invite honest feedback] + +- What's your initial reaction? +- How does this compare to what you do today? +- What would prevent you from using this? +- How much would this be worth to you? +- Who else would need to approve this purchase? +``` + +#### Phase 4: Wrap-up (5 min) +``` +"This has been incredibly helpful..." + +- Anything else I should have asked? +- Who else should I talk to about this? +- Can I follow up if I have more questions? +``` + +**Interview Best Practices:** +- Never ask "would you use this?" (people lie about future behavior) +- Ask about past behavior: "Tell me about the last time..." +- Embrace silence - count to 7 before filling gaps +- Watch for emotional reactions (pain = opportunity) +- Record with permission; take minimal notes during + +--- + +### Hypothesis Template + +**Format:** +``` +We believe that [building this feature/making this change] +For [target user segment] +Will [achieve this measurable outcome] + +We'll know we're right when [specific metric moves by X%] + +We'll know we're wrong when [falsification criteria] +``` + +**Example:** +``` +We believe that adding saved payment methods +For returning customers +Will increase checkout completion rate + +We'll know we're right when checkout completion increases by 15% + +We'll know we're wrong when completion rate stays flat after 2 weeks +or saved payment adoption is < 20% +``` + +**Hypothesis Quality Checklist:** +- [ ] Specific user segment defined +- [ ] Measurable outcome (number, not "better") +- [ ] Timeframe for measurement +- [ ] Clear falsification criteria +- [ ] Based on evidence (interviews, data) + +--- + +### Opportunity Solution Tree + +**Structure:** +``` +[DESIRED OUTCOME] + │ + ├── Opportunity 1: [User problem/need] + │ ├── Solution A + │ ├── Solution B + │ └── Experiment: [Test to validate] + │ + ├── Opportunity 2: [User problem/need] + │ ├── Solution C + │ └── Solution D + │ + └── Opportunity 3: [User problem/need] + └── Solution E +``` + +**Example:** +``` +[Increase monthly active users by 20%] + │ + ├── Users forget to return + │ ├── Weekly email digest + │ ├── Mobile push notifications + │ └── Test: A/B email frequency + │ + ├── New users don't find value quickly + │ ├── Improved onboarding wizard + │ └── Personalized first experience + │ + └── Users churn after free trial + ├── Extended trial for engaged users + └── Friction audit of upgrade flow +``` + +**Process:** +1. Start with measurable outcome (not solution) +2. Map opportunities from user research +3. Generate multiple solutions per opportunity +4. Design small experiments to validate +5. Prioritize based on learning potential + +--- + +### Jobs to Be Done + +**JTBD Statement Format:** +``` +When [situation/trigger] +I want to [motivation/job] +So I can [expected outcome] +``` + +**Example:** +``` +When I'm running late for a meeting +I want to notify attendees quickly +So I can set appropriate expectations and reduce anxiety +``` + +**Force Diagram:** +``` + ┌─────────────────┐ + Push from │ │ Pull toward + current ──────>│ SWITCH │<────── new + solution │ DECISION │ solution + │ │ + └─────────────────┘ + ^ ^ + | | + Anxiety of | | Habit of + change ──────┘ └────── status quo +``` + +**Interview Questions for JTBD:** +- When did you first realize you needed something like this? +- What were you using before? Why did you switch? +- What almost prevented you from switching? +- What would make you go back to the old way? + +--- + +## Metrics Frameworks + +### North Star Metric Framework + +**Criteria for a Good NSM:** +1. **Measures value delivery**: Captures what users get from product +2. **Leading indicator**: Predicts business success +3. **Actionable**: Teams can influence it +4. **Measurable**: Trackable on regular cadence + +**Examples by Business Type:** + +| Business | North Star Metric | Why | +|----------|-------------------|-----| +| Spotify | Time spent listening | Measures engagement value | +| Airbnb | Nights booked | Core transaction metric | +| Slack | Messages sent in channels | Team collaboration value | +| Dropbox | Files stored/synced | Storage utility delivered | +| Netflix | Hours watched | Entertainment value | + +**Supporting Metrics Structure:** +``` +[NORTH STAR METRIC] + │ + ├── Breadth: How many users? + ├── Depth: How engaged are they? + └── Frequency: How often do they engage? +``` + +--- + +### HEART Framework + +| Metric | Definition | Example Signals | +|--------|------------|-----------------| +| **Happiness** | Subjective satisfaction | NPS, CSAT, survey scores | +| **Engagement** | Depth of involvement | Session length, actions/session | +| **Adoption** | New user behavior | Signups, feature activation | +| **Retention** | Continued usage | D7/D30 retention, churn rate | +| **Task Success** | Efficiency & effectiveness | Completion rate, time-on-task, errors | + +**Goals-Signals-Metrics Process:** +1. **Goal**: What user behavior indicates success? +2. **Signal**: How would success manifest in data? +3. **Metric**: How do we measure the signal? + +**Example:** +``` +Feature: New checkout flow + +Goal: Users complete purchases faster +Signal: Reduced time in checkout, fewer drop-offs +Metrics: + - Median checkout time (target: <2 min) + - Checkout completion rate (target: 85%) + - Error rate (target: <2%) +``` + +--- + +### Funnel Analysis Template + +**Standard Funnel:** +``` +Acquisition → Activation → Retention → Revenue → Referral + │ │ │ │ │ + │ │ │ │ │ + How do First Come back Pay for Tell + they find "aha" regularly value others + you? moment +``` + +**Metrics per Stage:** + +| Stage | Key Metrics | Typical Benchmark | +|-------|-------------|-------------------| +| **Acquisition** | Visitors, CAC, channel mix | Varies by channel | +| **Activation** | Signup rate, onboarding completion | 20-30% visitor→signup | +| **Retention** | D1/D7/D30 retention, churn | D1: 40%, D7: 20%, D30: 10% | +| **Revenue** | Conversion rate, ARPU, LTV | 2-5% free→paid | +| **Referral** | NPS, viral coefficient, referrals/user | NPS > 50 is excellent | + +**Analysis Framework:** +1. Map current conversion rates at each stage +2. Identify biggest drop-off point +3. Qualitative research: Why are users leaving? +4. Hypothesis: What would improve conversion? +5. Test and measure + +--- + +### Feature Success Metrics + +| Metric | Definition | Target Range | +|--------|------------|--------------| +| **Adoption** | % users who try feature | 30-50% within 30 days | +| **Activation** | % who complete core action | 60-80% of adopters | +| **Frequency** | Uses per user per time | Weekly for engagement features | +| **Depth** | % of feature capability used | 50%+ of core functionality | +| **Retention** | Continued usage over time | 70%+ at 30 days | +| **Satisfaction** | Feature-specific NPS/rating | NPS > 30, Rating > 4.0 | + +**Measurement Cadence:** +- **Week 1**: Adoption and initial activation +- **Week 4**: Retention and depth +- **Week 8**: Long-term satisfaction and business impact + +--- + +## Strategic Frameworks + +### Product Vision Template + +**Format:** +``` +FOR [target customer] +WHO [statement of need or opportunity] +THE [product name] IS A [product category] +THAT [key benefit, compelling reason to use] +UNLIKE [primary competitive alternative] +OUR PRODUCT [statement of primary differentiation] +``` + +**Example:** +``` +FOR busy professionals +WHO need to stay informed without information overload +Briefme IS A personalized news digest +THAT delivers only relevant stories in 5 minutes +UNLIKE traditional news apps that require active browsing +OUR PRODUCT learns your interests and filters automatically +``` + +--- + +### Competitive Analysis Framework + +| Dimension | Us | Competitor A | Competitor B | +|-----------|----|--------------|--------------| +| **Target User** | | | | +| **Core Value Prop** | | | | +| **Pricing** | | | | +| **Key Features** | | | | +| **Strengths** | | | | +| **Weaknesses** | | | | +| **Market Position** | | | | + +**Strategic Questions:** +1. Where do we have parity? (table stakes) +2. Where do we differentiate? (competitive advantage) +3. Where are we behind? (gaps to close or ignore) +4. What can only we do? (unique capabilities) + +--- + +### Go-to-Market Checklist + +**Pre-Launch (4 weeks before):** +- [ ] Success metrics defined and instrumented +- [ ] Launch/rollback criteria established +- [ ] Support documentation ready +- [ ] Sales enablement materials complete +- [ ] Marketing assets prepared +- [ ] Beta feedback incorporated + +**Launch Week:** +- [ ] Staged rollout plan (1% → 10% → 50% → 100%) +- [ ] Monitoring dashboards live +- [ ] On-call rotation scheduled +- [ ] Communications ready (in-app, email, blog) +- [ ] Support team briefed + +**Post-Launch (2 weeks after):** +- [ ] Metrics review vs. targets +- [ ] User feedback synthesized +- [ ] Bug/issue triage complete +- [ ] Iteration plan defined +- [ ] Stakeholder update sent + +--- + +## Framework Selection Guide + +| Situation | Recommended Framework | +|-----------|----------------------| +| Quarterly roadmap planning | RICE + Portfolio Matrix | +| Sprint-level prioritization | MoSCoW | +| Quick feature comparison | ICE | +| Understanding user satisfaction | Kano | +| User research synthesis | JTBD + Opportunity Tree | +| Feature experiment design | Hypothesis Template | +| Success measurement | HEART + Feature Metrics | +| Strategy communication | North Star + Vision | + +--- + +*Last Updated: January 2025*