feat: add fact-checker skill v1.0.0

Add comprehensive fact-checking skill that verifies claims using web search
and official sources, then proposes corrections with user confirmation.

Features:
- 5-step workflow: identify → search → compare → report → apply
- Supports AI model specs, technical docs, statistics, general facts
- Source evaluation framework prioritizing official documentation
- Auto-correction with mandatory user approval gates
- Temporal context to prevent information decay

Real-world usage:
- Successfully updated AI model specs (Claude, GPT, Gemini)
- Corrected outdated version numbers and context windows
- Added temporal markers for time-sensitive information

Marketplace updates:
- Bumped version to 1.19.0
- Added fact-checker to plugins list
- Updated metadata description

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude Sonnet 4.5 <noreply@anthropic.com>
This commit is contained in:
daymade
2026-01-05 23:31:44 +08:00
parent 80cd0fe7e1
commit 7ba893d837
4 changed files with 487 additions and 2 deletions

View File

@@ -0,0 +1,4 @@
Security scan passed
Scanned at: 2026-01-05T23:29:12.688630
Tool: gitleaks + pattern-based validation
Content hash: f0ee8b5411aeb2d3058383ae536393e4e71e0fe0c240d0798a50aa10399870d4

188
fact-checker/README.md Normal file
View File

@@ -0,0 +1,188 @@
# Fact Checker
Verify factual claims in documents using web search and official sources, then apply corrections with user confirmation.
## Features
- ✅ Comprehensive fact verification across multiple domains
- 🔍 Searches authoritative sources (official docs, API specs, academic papers)
- 📊 Generates detailed correction reports with sources
- 🤖 Auto-applies corrections after user approval
- 🕐 Adds temporal context to prevent information decay
## Supported Claim Types
- **AI Model Specifications**: Context windows, pricing, features, benchmarks
- **Technical Documentation**: API capabilities, version numbers, library features
- **Statistical Data**: Metrics, benchmark scores, performance data
- **General Facts**: Any verifiable factual statement
## Usage Examples
### Example 1: Update Outdated AI Model Info
```
User: Fact-check the AI model specifications in section 2.1
```
**What happens:**
1. Identifies claims: "Claude 3.5 Sonnet: 200K tokens", "GPT-4o: 128K tokens"
2. Searches official documentation for current models
3. Finds: Claude Sonnet 4.5, GPT-5.2 with updated specs
4. Generates correction report with sources
5. Applies fixes after user confirms
### Example 2: Verify Technical Claims
```
User: Check if these library versions are still current
```
**What happens:**
1. Extracts version numbers from document
2. Checks package registries (npm, PyPI, etc.)
3. Identifies outdated versions
4. Suggests updates with changelog references
### Example 3: Validate Statistics
```
User: Verify the benchmark scores in this section
```
**What happens:**
1. Identifies numerical claims and metrics
2. Searches official benchmark publications
3. Compares document values vs. source data
4. Flags discrepancies with authoritative links
## Workflow
The skill follows a 5-step process:
```
Fact-checking Progress:
- [ ] Step 1: Identify factual claims
- [ ] Step 2: Search authoritative sources
- [ ] Step 3: Compare claims against sources
- [ ] Step 4: Generate correction report
- [ ] Step 5: Apply corrections with user approval
```
## Source Evaluation
**Preferred sources (in order):**
1. Official product pages and documentation
2. API documentation and developer guides
3. Official blog announcements
4. GitHub releases (for open source)
**Use with caution:**
- Third-party aggregators (verify against official sources)
- Blog posts and articles (cross-reference)
**Avoid:**
- Outdated documentation
- Unofficial wikis without citations
- Speculation and rumors
## Real-World Example
**Before:**
```markdown
AI 大模型的"上下文窗口"不断升级:
- Claude 3.5 Sonnet: 200K tokens约 15 万汉字)
- GPT-4o: 128K tokens约 10 万汉字)
- Gemini 1.5 Pro: 2M tokens约 150 万汉字)
```
**After fact-checking:**
```markdown
AI 大模型的"上下文窗口"不断升级(截至 2026 年 1 月):
- Claude Sonnet 4.5: 200K tokens约 15 万汉字)
- GPT-5.2: 400K tokens约 30 万汉字)
- Gemini 3 Pro: 1M tokens约 75 万汉字)
```
**Changes made:**
- ✅ Updated Claude 3.5 Sonnet → Claude Sonnet 4.5
- ✅ Corrected GPT-4o (128K) → GPT-5.2 (400K)
- ✅ Fixed Gemini 1.5 Pro (2M) → Gemini 3 Pro (1M)
- ✅ Added temporal marker "截至 2026 年 1 月"
## Installation
```bash
# Via CCPM (recommended)
ccpm install @daymade-skills/fact-checker
# Manual installation
Download fact-checker.zip and install through Claude Code
```
## Trigger Keywords
The skill activates when you mention:
- "fact-check this document"
- "verify these claims"
- "check if this is accurate"
- "update outdated information"
- "validate the data"
## Configuration
No configuration required. The skill works out of the box.
## Limitations
**Cannot verify:**
- Subjective opinions or judgments
- Future predictions or specifications
- Claims requiring paywalled sources
- Disputed facts without authoritative consensus
**For such cases**, the skill will:
- Note the limitation in the report
- Suggest qualification language
- Recommend user research or expert consultation
## Best Practices
### For Authors
1. **Run regularly**: Fact-check documents periodically to catch outdated info
2. **Include dates**: Add temporal markers like "as of [date]" to claims
3. **Cite sources**: Keep original source links for future verification
4. **Review reports**: Always review the correction report before applying changes
### For Fact-Checking
1. **Be specific**: Target specific sections rather than entire books
2. **Verify critical claims first**: Prioritize high-impact information
3. **Cross-reference**: For important claims, verify across multiple sources
4. **Update regularly**: Technical specs change frequently - recheck periodically
## Development
Created with skill-creator v1.2.2 following Anthropic's best practices.
**Testing:**
- Verified on Claude Sonnet 4.5, Opus 4.5, and Haiku 4
- Tested with real-world documentation updates
- Validated correction workflow with user approval gates
## Version History
### 1.0.0 (2026-01-05)
- Initial release
- Support for AI models, technical docs, statistics
- Auto-correction with user approval
- Comprehensive source evaluation framework
## License
MIT License - See repository for details
## Contributing
Issues and pull requests welcome at [daymade/claude-code-skills](https://github.com/daymade/claude-code-skills)

283
fact-checker/SKILL.md Normal file
View File

@@ -0,0 +1,283 @@
---
name: fact-checker
description: Verifies factual claims in documents using web search and official sources, then proposes corrections with user confirmation. Use when the user asks to fact-check, verify information, validate claims, check accuracy, or update outdated information in documents. Supports AI model specs, technical documentation, statistics, and general factual statements.
---
# Fact Checker
Verify factual claims in documents and propose corrections backed by authoritative sources.
## When to use
Trigger when users request:
- "Fact-check this document"
- "Verify these AI model specifications"
- "Check if this information is still accurate"
- "Update outdated data in this file"
- "Validate the claims in this section"
## Workflow
Copy this checklist to track progress:
```
Fact-checking Progress:
- [ ] Step 1: Identify factual claims
- [ ] Step 2: Search authoritative sources
- [ ] Step 3: Compare claims against sources
- [ ] Step 4: Generate correction report
- [ ] Step 5: Apply corrections with user approval
```
### Step 1: Identify factual claims
Scan the document for verifiable statements:
**Target claim types:**
- Technical specifications (context windows, pricing, features)
- Version numbers and release dates
- Statistical data and metrics
- API capabilities and limitations
- Benchmark scores and performance data
**Skip subjective content:**
- Opinions and recommendations
- Explanatory prose
- Tutorial instructions
- Architectural discussions
### Step 2: Search authoritative sources
For each claim, search official sources:
**AI models:**
- Official announcement pages (anthropic.com/news, openai.com/index, blog.google)
- API documentation (platform.claude.com/docs, platform.openai.com/docs)
- Developer guides and release notes
**Technical libraries:**
- Official documentation sites
- GitHub repositories (releases, README)
- Package registries (npm, PyPI, crates.io)
**General claims:**
- Academic papers and research
- Government statistics
- Industry standards bodies
**Search strategy:**
- Use model names + specification (e.g., "Claude Opus 4.5 context window")
- Include current year for recent information
- Verify from multiple sources when possible
### Step 3: Compare claims against sources
Create a comparison table:
| Claim in Document | Source Information | Status | Authoritative Source |
|-------------------|-------------------|--------|---------------------|
| Claude 3.5 Sonnet: 200K tokens | Claude Sonnet 4.5: 200K tokens | ❌ Outdated model name | platform.claude.com/docs |
| GPT-4o: 128K tokens | GPT-5.2: 400K tokens | ❌ Incorrect version & spec | openai.com/index/gpt-5-2 |
**Status codes:**
- ✅ Accurate - claim matches sources
- ❌ Incorrect - claim contradicts sources
- ⚠️ Outdated - claim was true but superseded
- ❓ Unverifiable - no authoritative source found
### Step 4: Generate correction report
Present findings in structured format:
```markdown
## Fact-Check Report
### Summary
- Total claims checked: X
- Accurate: Y
- Issues found: Z
### Issues Requiring Correction
#### Issue 1: Outdated AI Model Reference
**Location:** Line 77-80 in docs/file.md
**Current claim:** "Claude 3.5 Sonnet: 200K tokens"
**Correction:** "Claude Sonnet 4.5: 200K tokens"
**Source:** https://platform.claude.com/docs/en/build-with-claude/context-windows
**Rationale:** Claude 3.5 Sonnet has been superseded by Claude Sonnet 4.5 (released Sept 2025)
#### Issue 2: Incorrect Context Window
**Location:** Line 79 in docs/file.md
**Current claim:** "GPT-4o: 128K tokens"
**Correction:** "GPT-5.2: 400K tokens"
**Source:** https://openai.com/index/introducing-gpt-5-2/
**Rationale:** 128K was output limit; context window is 400K. Model also updated to GPT-5.2
```
### Step 5: Apply corrections with user approval
**Before making changes:**
1. Show the correction report to the user
2. Wait for explicit approval: "Should I apply these corrections?"
3. Only proceed after confirmation
**When applying corrections:**
```python
# Use Edit tool to update document
# Example:
Edit(
file_path="docs/03-写作规范/AI辅助写书方法论.md",
old_string="- Claude 3.5 Sonnet: 200K tokens约 15 万汉字)",
new_string="- Claude Sonnet 4.5: 200K tokens约 15 万汉字)"
)
```
**After corrections:**
1. Verify all edits were applied successfully
2. Note the correction summary (e.g., "Updated 4 claims in section 2.1")
3. Remind user to commit changes
## Search best practices
### Query construction
**Good queries** (specific, current):
- "Claude Opus 4.5 context window 2026"
- "GPT-5.2 official release announcement"
- "Gemini 3 Pro token limit specifications"
**Poor queries** (vague, generic):
- "Claude context"
- "AI models"
- "Latest version"
### Source evaluation
**Prefer official sources:**
1. Product official pages (highest authority)
2. API documentation
3. Official blog announcements
4. GitHub releases (for open source)
**Use with caution:**
- Third-party aggregators (llm-stats.com, etc.) - verify against official sources
- Blog posts and articles - cross-reference claims
- Social media - only for announcements, verify elsewhere
**Avoid:**
- Outdated documentation
- Unofficial wikis without citations
- Speculation and rumors
### Handling ambiguity
When sources conflict:
1. Prioritize most recent official documentation
2. Note the discrepancy in the report
3. Present both sources to the user
4. Recommend contacting vendor if critical
When no source found:
1. Mark as ❓ Unverifiable
2. Suggest alternative phrasing: "According to [Source] as of [Date]..."
3. Recommend adding qualification: "approximately", "reported as"
## Special considerations
### Time-sensitive information
Always include temporal context:
**Good corrections:**
- "截至 2026 年 1 月" (As of January 2026)
- "Claude Sonnet 4.5 (released September 2025)"
**Poor corrections:**
- "Latest version" (becomes outdated)
- "Current model" (ambiguous timeframe)
### Numerical precision
Match precision to source:
**Source says:** "approximately 1 million tokens"
**Write:** "1M tokens (approximately)"
**Source says:** "200,000 token context window"
**Write:** "200K tokens" (exact)
### Citation format
Include citations in corrections:
```markdown
> **注**:具体上下文窗口以模型官方文档为准,本书写作时使用 Claude Sonnet 4.5 为主要工具。
```
Link to sources when possible.
## Examples
### Example 1: Technical specification update
**User request:** "Fact-check the AI model context windows in section 2.1"
**Process:**
1. Identify claims: Claude 3.5 Sonnet (200K), GPT-4o (128K), Gemini 1.5 Pro (2M)
2. Search official docs for current models
3. Find: Claude Sonnet 4.5, GPT-5.2, Gemini 3 Pro
4. Generate report showing discrepancies
5. Apply corrections after approval
### Example 2: Statistical data verification
**User request:** "Verify the benchmark scores in chapter 5"
**Process:**
1. Extract numerical claims
2. Search for official benchmark publications
3. Compare reported vs. source values
4. Flag any discrepancies with source links
5. Update with verified figures
### Example 3: Version number validation
**User request:** "Check if these library versions are still current"
**Process:**
1. List all version numbers mentioned
2. Check package registries (npm, PyPI, etc.)
3. Identify outdated versions
4. Suggest updates with changelog references
5. Update after user confirms
## Quality checklist
Before completing fact-check:
- [ ] All factual claims identified and categorized
- [ ] Each claim verified against official sources
- [ ] Sources are authoritative and current
- [ ] Correction report is clear and actionable
- [ ] Temporal context included where relevant
- [ ] User approval obtained before changes
- [ ] All edits verified successful
- [ ] Summary provided to user
## Limitations
**This skill cannot:**
- Verify subjective opinions or judgments
- Access paywalled or restricted sources
- Determine "truth" in disputed claims
- Predict future specifications or features
**For such cases:**
- Note the limitation in the report
- Suggest qualification language
- Recommend user research or expert consultation