Update skill docs and resources
This commit is contained in:
231
product-analysis/SKILL.md
Normal file
231
product-analysis/SKILL.md
Normal file
@@ -0,0 +1,231 @@
|
||||
---
|
||||
name: product-analysis
|
||||
description: Multi-path parallel product analysis with cross-model test-time compute scaling. Spawns parallel agents (Claude Code agent teams + Codex CLI) to explore product from multiple perspectives, then synthesizes findings into actionable optimization plans. Can invoke competitors-analysis for competitive benchmarking. Use when "product audit", "self-review", "发布前审查", "产品分析", "analyze our product", "UX audit", or "信息架构审计".
|
||||
argument-hint: [scope: full|ux|api|arch|compare]
|
||||
---
|
||||
|
||||
# Product Analysis
|
||||
|
||||
Multi-path parallel product analysis that combines **Claude Code agent teams** and **Codex CLI** for cross-model test-time compute scaling.
|
||||
|
||||
**Core principle**: Same analysis task, multiple AI perspectives, deep synthesis.
|
||||
|
||||
## How It Works
|
||||
|
||||
```
|
||||
/product-analysis full
|
||||
│
|
||||
├─ Step 0: Auto-detect available tools (codex? competitors?)
|
||||
│
|
||||
┌────┼──────────────┐
|
||||
│ │ │
|
||||
Claude Code Codex CLI (auto-detected)
|
||||
Task Agents (background Bash)
|
||||
(Explore ×3-5) (×2-3 parallel)
|
||||
│ │
|
||||
└────────┬──────────┘
|
||||
│
|
||||
Synthesis (main context)
|
||||
│
|
||||
Structured Report
|
||||
```
|
||||
|
||||
## Step 0: Auto-Detect Available Tools
|
||||
|
||||
Before launching any agents, detect what tools are available:
|
||||
|
||||
```bash
|
||||
# Check if Codex CLI is installed
|
||||
which codex 2>/dev/null && codex --version
|
||||
```
|
||||
|
||||
**Decision logic**:
|
||||
- If `codex` is found: Inform the user — "Codex CLI detected (version X). Will run cross-model analysis for richer perspectives."
|
||||
- If `codex` is not found: Silently proceed with Claude Code agents only. Do NOT ask the user to install anything.
|
||||
|
||||
Also detect the project type to tailor agent prompts:
|
||||
```bash
|
||||
# Detect project type
|
||||
ls package.json 2>/dev/null # Node.js/React
|
||||
ls pyproject.toml 2>/dev/null # Python
|
||||
ls Cargo.toml 2>/dev/null # Rust
|
||||
ls go.mod 2>/dev/null # Go
|
||||
```
|
||||
|
||||
## Scope Modes
|
||||
|
||||
Parse `$ARGUMENTS` to determine analysis scope:
|
||||
|
||||
| Scope | What it covers | Typical agents |
|
||||
|-------|---------------|----------------|
|
||||
| `full` | UX + API + Architecture + Docs (default) | 5 Claude + Codex (if available) |
|
||||
| `ux` | Frontend navigation, information density, user journey, empty state, onboarding | 3 Claude + Codex (if available) |
|
||||
| `api` | Backend API coverage, endpoint health, error handling, consistency | 2 Claude + Codex (if available) |
|
||||
| `arch` | Module structure, dependency graph, code duplication, separation of concerns | 2 Claude + Codex (if available) |
|
||||
| `compare X Y` | Self-audit + competitive benchmarking (invokes `/competitors-analysis`) | 3 Claude + competitors-analysis |
|
||||
|
||||
## Phase 1: Parallel Exploration
|
||||
|
||||
Launch all exploration agents simultaneously using Task tool (background mode).
|
||||
|
||||
### Claude Code Agents (always)
|
||||
|
||||
For each dimension, spawn a Task agent with `subagent_type: Explore` and `run_in_background: true`:
|
||||
|
||||
**Agent A — Frontend Navigation & Information Density**
|
||||
```
|
||||
Explore the frontend navigation structure and entry points:
|
||||
1. App.tsx: How many top-level components are mounted simultaneously?
|
||||
2. Left sidebar: How many buttons/entries? What does each link to?
|
||||
3. Right sidebar: How many tabs? How many sections per tab?
|
||||
4. Floating panels: How many drawers/modals? Which overlap in functionality?
|
||||
5. Count total first-screen interactive elements for a new user.
|
||||
6. Identify duplicate entry points (same feature accessible from 2+ places).
|
||||
Give specific file paths, line numbers, and element counts.
|
||||
```
|
||||
|
||||
**Agent B — User Journey & Empty State**
|
||||
```
|
||||
Explore the new user experience:
|
||||
1. Empty state page: What does a user with no sessions see? Count clickable elements.
|
||||
2. Onboarding flow: How many steps? What information is presented?
|
||||
3. Prompt input area: How many buttons/controls surround the input box? Which are high-frequency vs low-frequency?
|
||||
4. Mobile adaptation: How many nav items? How does it differ from desktop?
|
||||
5. Estimate: Can a new user complete their first conversation in 3 minutes?
|
||||
Give specific file paths, line numbers, and UX assessment.
|
||||
```
|
||||
|
||||
**Agent C — Backend API & Health**
|
||||
```
|
||||
Explore the backend API surface:
|
||||
1. List ALL API endpoints (method + path + purpose).
|
||||
2. Identify endpoints that are unused or have no frontend consumer.
|
||||
3. Check error handling consistency (do all endpoints return structured errors?).
|
||||
4. Check authentication/authorization patterns (which endpoints require auth?).
|
||||
5. Identify any endpoints that duplicate functionality.
|
||||
Give specific file paths and line numbers.
|
||||
```
|
||||
|
||||
**Agent D — Architecture & Module Structure** (full/arch scope only)
|
||||
```
|
||||
Explore the module structure and dependencies:
|
||||
1. Map the module dependency graph (which modules import which).
|
||||
2. Identify circular dependencies or tight coupling.
|
||||
3. Find code duplication across modules (same pattern in 3+ places).
|
||||
4. Check separation of concerns (does each module have a single responsibility?).
|
||||
5. Identify dead code or unused exports.
|
||||
Give specific file paths and line numbers.
|
||||
```
|
||||
|
||||
**Agent E — Documentation & Config Consistency** (full scope only)
|
||||
```
|
||||
Explore documentation and configuration:
|
||||
1. Compare README claims vs actual implemented features.
|
||||
2. Check config file consistency (base.yaml vs .env.example vs code defaults).
|
||||
3. Find outdated documentation (references to removed features/files).
|
||||
4. Check test coverage gaps (which modules have no tests?).
|
||||
Give specific file paths and line numbers.
|
||||
```
|
||||
|
||||
### Codex CLI Agents (auto-detected)
|
||||
|
||||
If Codex CLI was detected in Step 0, launch parallel Codex analyses via background Bash.
|
||||
|
||||
Each Codex invocation gets the same dimensional prompt but from a different model's perspective:
|
||||
|
||||
```bash
|
||||
codex -m o4-mini \
|
||||
-c model_reasoning_effort="high" \
|
||||
--full-auto \
|
||||
"Analyze the frontend navigation structure of this project. Count all interactive elements visible to a new user on first screen. Identify duplicate entry points where the same feature is accessible from 2+ places. Give specific file paths and counts."
|
||||
```
|
||||
|
||||
Run 2-3 Codex commands in parallel (background Bash), one per major dimension.
|
||||
|
||||
**Important**: Codex runs in the project's working directory. It has full filesystem access. The `--full-auto` flag (or `--dangerously-bypass-approvals-and-sandbox` for older versions) enables autonomous execution.
|
||||
|
||||
## Phase 2: Competitive Benchmarking (compare scope only)
|
||||
|
||||
When scope is `compare`, invoke the competitors-analysis skill for each competitor:
|
||||
|
||||
```
|
||||
Use the Skill tool to invoke: /competitors-analysis {competitor-name} {competitor-url}
|
||||
```
|
||||
|
||||
This delegates to the orthogonal `competitors-analysis` skill which handles:
|
||||
- Repository cloning and validation
|
||||
- Evidence-based code analysis (file:line citations)
|
||||
- Competitor profile generation
|
||||
|
||||
## Phase 3: Synthesis
|
||||
|
||||
After all agents complete, synthesize findings in the main conversation context.
|
||||
|
||||
### Cross-Validation
|
||||
|
||||
Compare findings across agents (Claude vs Claude, Claude vs Codex):
|
||||
- **Agreement** = high confidence finding
|
||||
- **Disagreement** = investigate deeper (one agent may have missed context)
|
||||
- **Codex-only finding** = different model perspective, validate manually
|
||||
|
||||
### Quantification
|
||||
|
||||
Extract hard numbers from agent reports:
|
||||
|
||||
| Metric | What to measure |
|
||||
|--------|----------------|
|
||||
| First-screen interactive elements | Total count of buttons/links/inputs visible to new user |
|
||||
| Feature entry point duplication | Number of features with 2+ entry points |
|
||||
| API endpoints without frontend consumer | Count of unused backend routes |
|
||||
| Onboarding steps to first value | Steps from launch to first successful action |
|
||||
| Module coupling score | Number of circular or bi-directional dependencies |
|
||||
|
||||
### Structured Output
|
||||
|
||||
Produce a layered optimization report:
|
||||
|
||||
```markdown
|
||||
## Product Analysis Report
|
||||
|
||||
### Executive Summary
|
||||
[1-2 sentences: key finding]
|
||||
|
||||
### Quantified Findings
|
||||
| Metric | Value | Assessment |
|
||||
|--------|-------|------------|
|
||||
| ... | ... | ... |
|
||||
|
||||
### P0: Critical (block launch)
|
||||
[Issues that prevent basic usability]
|
||||
|
||||
### P1: High Priority (launch week)
|
||||
[Issues that significantly degrade experience]
|
||||
|
||||
### P2: Medium Priority (next sprint)
|
||||
[Issues worth addressing but not blocking]
|
||||
|
||||
### Cross-Model Insights
|
||||
[Findings that only one model identified — worth investigating]
|
||||
|
||||
### Competitive Position (if compare scope)
|
||||
[How we compare on key dimensions]
|
||||
```
|
||||
|
||||
## Workflow Checklist
|
||||
|
||||
- [ ] Parse `$ARGUMENTS` for scope
|
||||
- [ ] Auto-detect Codex CLI availability (`which codex`)
|
||||
- [ ] Auto-detect project type (package.json / pyproject.toml / etc.)
|
||||
- [ ] Launch Claude Code Explore agents (3-5 parallel, background)
|
||||
- [ ] Launch Codex CLI commands (2-3 parallel, background) if detected
|
||||
- [ ] Invoke `/competitors-analysis` if `compare` scope
|
||||
- [ ] Collect all agent results
|
||||
- [ ] Cross-validate findings
|
||||
- [ ] Quantify metrics
|
||||
- [ ] Generate structured report with P0/P1/P2 priorities
|
||||
|
||||
## References
|
||||
|
||||
- [references/analysis_dimensions.md](references/analysis_dimensions.md) — Detailed audit dimension definitions and prompts
|
||||
- [references/synthesis_methodology.md](references/synthesis_methodology.md) — How to weight and merge multi-agent findings
|
||||
- [references/codex_patterns.md](references/codex_patterns.md) — Codex CLI invocation patterns and flag reference
|
||||
Reference in New Issue
Block a user