Files
2026-02-23 16:16:58 +08:00

69 lines
2.5 KiB
Markdown

# Synthesis Methodology
How to weight, merge, and validate findings from multiple parallel agents.
## Multi-Agent Synthesis Framework
### Step 1: Collect Raw Findings
Wait for all agents to complete. For each agent, extract:
- **Quantitative data**: counts, measurements, lists
- **Qualitative assessments**: good/bad/unclear judgments
- **Evidence**: file paths, line numbers, code snippets
### Step 2: Cross-Validation Matrix
Create a matrix comparing findings across agents:
```
| Finding | Agent A | Agent B | Codex | Confidence |
|---------|---------|---------|-------|------------|
| "57 interactive elements on first screen" | 57 | 54 | 61 | HIGH (3/3 agree on magnitude) |
| "Skills has 3 entry points" | 3 | 3 | 2 | HIGH (2/3 exact match) |
| "Risk pages should be removed" | Yes | - | No | LOW (disagreement, investigate) |
```
**Confidence levels**:
- **HIGH**: 2+ agents agree (exact or same magnitude)
- **MEDIUM**: 1 agent found, others didn't look
- **LOW**: Agents disagree — requires manual investigation
### Step 3: Disagreement Resolution
When agents disagree:
1. Check if they analyzed different files/scopes
2. Check if one agent missed context (e.g., conditional rendering)
3. If genuine disagreement, note both perspectives in report
4. Codex-only findings are "different model perspective" — valuable but need validation
### Step 4: Priority Assignment
**P0 (Critical)**: Issues that prevent a new user from completing basic tasks
- Examples: broken onboarding, missing error messages, dead navigation links
**P1 (High)**: Issues that significantly increase cognitive load or confusion
- Examples: duplicate entry points, information overload, unclear primary action
**P2 (Medium)**: Issues worth addressing but not blocking launch
- Examples: unused API endpoints, minor inconsistencies, missing edge case handling
### Step 5: Report Generation
Structure the report for actionability:
1. **Executive Summary** (2-3 sentences, the "so what")
2. **Quantified Metrics** (hard numbers, no adjectives)
3. **P0 Issues** (with specific file:line references)
4. **P1 Issues** (with suggested fixes)
5. **P2 Issues** (backlog items)
6. **Cross-Model Insights** (findings unique to one model)
7. **Competitive Position** (if compare scope was used)
## Weighting Rules
- Quantitative findings (counts, measurements) > Qualitative judgments
- Code-evidenced findings > Assumption-based findings
- Multi-agent agreement > Single-agent finding
- User-facing issues > Internal code quality issues
- Findings with clear fix path > Vague "should improve" suggestions