feat: add fact-checker skill v1.0.0

Add comprehensive fact-checking skill that verifies claims using web search
and official sources, then proposes corrections with user confirmation.

Features:
- 5-step workflow: identify → search → compare → report → apply
- Supports AI model specs, technical docs, statistics, general facts
- Source evaluation framework prioritizing official documentation
- Auto-correction with mandatory user approval gates
- Temporal context to prevent information decay

Real-world usage:
- Successfully updated AI model specs (Claude, GPT, Gemini)
- Corrected outdated version numbers and context windows
- Added temporal markers for time-sensitive information

Marketplace updates:
- Bumped version to 1.19.0
- Added fact-checker to plugins list
- Updated metadata description

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude Sonnet 4.5 <noreply@anthropic.com>
This commit is contained in:
daymade
2026-01-05 23:31:44 +08:00
parent 80cd0fe7e1
commit 7ba893d837
4 changed files with 487 additions and 2 deletions

View File

@@ -5,8 +5,8 @@
"email": "daymadev89@gmail.com"
},
"metadata": {
"description": "Professional Claude Code skills for GitHub operations, document conversion, diagram generation, statusline customization, Teams communication, repomix utilities, skill creation, CLI demo generation, LLM icon access, Cloudflare troubleshooting, UI design system extraction, professional presentation creation, YouTube video downloading, secure repomix packaging, ASR transcription correction, video comparison quality analysis, comprehensive QA testing infrastructure, prompt optimization with EARS methodology, session history recovery, documentation cleanup, PDF generation with Chinese font support, CLAUDE.md progressive disclosure optimization, CCPM skill registry search and management, Promptfoo LLM evaluation framework, and iOS app development with XcodeGen and SwiftUI",
"version": "1.18.1",
"description": "Professional Claude Code skills for GitHub operations, document conversion, diagram generation, statusline customization, Teams communication, repomix utilities, skill creation, CLI demo generation, LLM icon access, Cloudflare troubleshooting, UI design system extraction, professional presentation creation, YouTube video downloading, secure repomix packaging, ASR transcription correction, video comparison quality analysis, comprehensive QA testing infrastructure, prompt optimization with EARS methodology, session history recovery, documentation cleanup, PDF generation with Chinese font support, CLAUDE.md progressive disclosure optimization, CCPM skill registry search and management, Promptfoo LLM evaluation framework, iOS app development with XcodeGen and SwiftUI, and fact-checking with automated corrections",
"version": "1.19.0",
"homepage": "https://github.com/daymade/claude-code-skills"
},
"plugins": [
@@ -259,6 +259,16 @@
"category": "developer-tools",
"keywords": ["ios", "xcodegen", "swiftui", "spm", "avfoundation", "camera", "code-signing", "device-deployment", "swift", "xcode", "testing"],
"skills": ["./iOS-APP-developer"]
},
{
"name": "fact-checker",
"description": "Verifies factual claims in documents using web search and official sources, then proposes corrections with user confirmation. Use when the user asks to fact-check, verify information, validate claims, check accuracy, or update outdated information in documents. Supports AI model specs, technical documentation, statistics, and general factual statements",
"source": "./",
"strict": false,
"version": "1.0.0",
"category": "productivity",
"keywords": ["fact-checking", "verification", "accuracy", "sources", "validation", "corrections", "web-search"],
"skills": ["./fact-checker"]
}
]
}

View File

@@ -0,0 +1,4 @@
Security scan passed
Scanned at: 2026-01-05T23:29:12.688630
Tool: gitleaks + pattern-based validation
Content hash: f0ee8b5411aeb2d3058383ae536393e4e71e0fe0c240d0798a50aa10399870d4

188
fact-checker/README.md Normal file
View File

@@ -0,0 +1,188 @@
# Fact Checker
Verify factual claims in documents using web search and official sources, then apply corrections with user confirmation.
## Features
- ✅ Comprehensive fact verification across multiple domains
- 🔍 Searches authoritative sources (official docs, API specs, academic papers)
- 📊 Generates detailed correction reports with sources
- 🤖 Auto-applies corrections after user approval
- 🕐 Adds temporal context to prevent information decay
## Supported Claim Types
- **AI Model Specifications**: Context windows, pricing, features, benchmarks
- **Technical Documentation**: API capabilities, version numbers, library features
- **Statistical Data**: Metrics, benchmark scores, performance data
- **General Facts**: Any verifiable factual statement
## Usage Examples
### Example 1: Update Outdated AI Model Info
```
User: Fact-check the AI model specifications in section 2.1
```
**What happens:**
1. Identifies claims: "Claude 3.5 Sonnet: 200K tokens", "GPT-4o: 128K tokens"
2. Searches official documentation for current models
3. Finds: Claude Sonnet 4.5, GPT-5.2 with updated specs
4. Generates correction report with sources
5. Applies fixes after user confirms
### Example 2: Verify Technical Claims
```
User: Check if these library versions are still current
```
**What happens:**
1. Extracts version numbers from document
2. Checks package registries (npm, PyPI, etc.)
3. Identifies outdated versions
4. Suggests updates with changelog references
### Example 3: Validate Statistics
```
User: Verify the benchmark scores in this section
```
**What happens:**
1. Identifies numerical claims and metrics
2. Searches official benchmark publications
3. Compares document values vs. source data
4. Flags discrepancies with authoritative links
## Workflow
The skill follows a 5-step process:
```
Fact-checking Progress:
- [ ] Step 1: Identify factual claims
- [ ] Step 2: Search authoritative sources
- [ ] Step 3: Compare claims against sources
- [ ] Step 4: Generate correction report
- [ ] Step 5: Apply corrections with user approval
```
## Source Evaluation
**Preferred sources (in order):**
1. Official product pages and documentation
2. API documentation and developer guides
3. Official blog announcements
4. GitHub releases (for open source)
**Use with caution:**
- Third-party aggregators (verify against official sources)
- Blog posts and articles (cross-reference)
**Avoid:**
- Outdated documentation
- Unofficial wikis without citations
- Speculation and rumors
## Real-World Example
**Before:**
```markdown
AI 大模型的"上下文窗口"不断升级:
- Claude 3.5 Sonnet: 200K tokens约 15 万汉字)
- GPT-4o: 128K tokens约 10 万汉字)
- Gemini 1.5 Pro: 2M tokens约 150 万汉字)
```
**After fact-checking:**
```markdown
AI 大模型的"上下文窗口"不断升级(截至 2026 年 1 月):
- Claude Sonnet 4.5: 200K tokens约 15 万汉字)
- GPT-5.2: 400K tokens约 30 万汉字)
- Gemini 3 Pro: 1M tokens约 75 万汉字)
```
**Changes made:**
- ✅ Updated Claude 3.5 Sonnet → Claude Sonnet 4.5
- ✅ Corrected GPT-4o (128K) → GPT-5.2 (400K)
- ✅ Fixed Gemini 1.5 Pro (2M) → Gemini 3 Pro (1M)
- ✅ Added temporal marker "截至 2026 年 1 月"
## Installation
```bash
# Via CCPM (recommended)
ccpm install @daymade-skills/fact-checker
# Manual installation
Download fact-checker.zip and install through Claude Code
```
## Trigger Keywords
The skill activates when you mention:
- "fact-check this document"
- "verify these claims"
- "check if this is accurate"
- "update outdated information"
- "validate the data"
## Configuration
No configuration required. The skill works out of the box.
## Limitations
**Cannot verify:**
- Subjective opinions or judgments
- Future predictions or specifications
- Claims requiring paywalled sources
- Disputed facts without authoritative consensus
**For such cases**, the skill will:
- Note the limitation in the report
- Suggest qualification language
- Recommend user research or expert consultation
## Best Practices
### For Authors
1. **Run regularly**: Fact-check documents periodically to catch outdated info
2. **Include dates**: Add temporal markers like "as of [date]" to claims
3. **Cite sources**: Keep original source links for future verification
4. **Review reports**: Always review the correction report before applying changes
### For Fact-Checking
1. **Be specific**: Target specific sections rather than entire books
2. **Verify critical claims first**: Prioritize high-impact information
3. **Cross-reference**: For important claims, verify across multiple sources
4. **Update regularly**: Technical specs change frequently - recheck periodically
## Development
Created with skill-creator v1.2.2 following Anthropic's best practices.
**Testing:**
- Verified on Claude Sonnet 4.5, Opus 4.5, and Haiku 4
- Tested with real-world documentation updates
- Validated correction workflow with user approval gates
## Version History
### 1.0.0 (2026-01-05)
- Initial release
- Support for AI models, technical docs, statistics
- Auto-correction with user approval
- Comprehensive source evaluation framework
## License
MIT License - See repository for details
## Contributing
Issues and pull requests welcome at [daymade/claude-code-skills](https://github.com/daymade/claude-code-skills)

283
fact-checker/SKILL.md Normal file
View File

@@ -0,0 +1,283 @@
---
name: fact-checker
description: Verifies factual claims in documents using web search and official sources, then proposes corrections with user confirmation. Use when the user asks to fact-check, verify information, validate claims, check accuracy, or update outdated information in documents. Supports AI model specs, technical documentation, statistics, and general factual statements.
---
# Fact Checker
Verify factual claims in documents and propose corrections backed by authoritative sources.
## When to use
Trigger when users request:
- "Fact-check this document"
- "Verify these AI model specifications"
- "Check if this information is still accurate"
- "Update outdated data in this file"
- "Validate the claims in this section"
## Workflow
Copy this checklist to track progress:
```
Fact-checking Progress:
- [ ] Step 1: Identify factual claims
- [ ] Step 2: Search authoritative sources
- [ ] Step 3: Compare claims against sources
- [ ] Step 4: Generate correction report
- [ ] Step 5: Apply corrections with user approval
```
### Step 1: Identify factual claims
Scan the document for verifiable statements:
**Target claim types:**
- Technical specifications (context windows, pricing, features)
- Version numbers and release dates
- Statistical data and metrics
- API capabilities and limitations
- Benchmark scores and performance data
**Skip subjective content:**
- Opinions and recommendations
- Explanatory prose
- Tutorial instructions
- Architectural discussions
### Step 2: Search authoritative sources
For each claim, search official sources:
**AI models:**
- Official announcement pages (anthropic.com/news, openai.com/index, blog.google)
- API documentation (platform.claude.com/docs, platform.openai.com/docs)
- Developer guides and release notes
**Technical libraries:**
- Official documentation sites
- GitHub repositories (releases, README)
- Package registries (npm, PyPI, crates.io)
**General claims:**
- Academic papers and research
- Government statistics
- Industry standards bodies
**Search strategy:**
- Use model names + specification (e.g., "Claude Opus 4.5 context window")
- Include current year for recent information
- Verify from multiple sources when possible
### Step 3: Compare claims against sources
Create a comparison table:
| Claim in Document | Source Information | Status | Authoritative Source |
|-------------------|-------------------|--------|---------------------|
| Claude 3.5 Sonnet: 200K tokens | Claude Sonnet 4.5: 200K tokens | ❌ Outdated model name | platform.claude.com/docs |
| GPT-4o: 128K tokens | GPT-5.2: 400K tokens | ❌ Incorrect version & spec | openai.com/index/gpt-5-2 |
**Status codes:**
- ✅ Accurate - claim matches sources
- ❌ Incorrect - claim contradicts sources
- ⚠️ Outdated - claim was true but superseded
- ❓ Unverifiable - no authoritative source found
### Step 4: Generate correction report
Present findings in structured format:
```markdown
## Fact-Check Report
### Summary
- Total claims checked: X
- Accurate: Y
- Issues found: Z
### Issues Requiring Correction
#### Issue 1: Outdated AI Model Reference
**Location:** Line 77-80 in docs/file.md
**Current claim:** "Claude 3.5 Sonnet: 200K tokens"
**Correction:** "Claude Sonnet 4.5: 200K tokens"
**Source:** https://platform.claude.com/docs/en/build-with-claude/context-windows
**Rationale:** Claude 3.5 Sonnet has been superseded by Claude Sonnet 4.5 (released Sept 2025)
#### Issue 2: Incorrect Context Window
**Location:** Line 79 in docs/file.md
**Current claim:** "GPT-4o: 128K tokens"
**Correction:** "GPT-5.2: 400K tokens"
**Source:** https://openai.com/index/introducing-gpt-5-2/
**Rationale:** 128K was output limit; context window is 400K. Model also updated to GPT-5.2
```
### Step 5: Apply corrections with user approval
**Before making changes:**
1. Show the correction report to the user
2. Wait for explicit approval: "Should I apply these corrections?"
3. Only proceed after confirmation
**When applying corrections:**
```python
# Use Edit tool to update document
# Example:
Edit(
file_path="docs/03-写作规范/AI辅助写书方法论.md",
old_string="- Claude 3.5 Sonnet: 200K tokens约 15 万汉字)",
new_string="- Claude Sonnet 4.5: 200K tokens约 15 万汉字)"
)
```
**After corrections:**
1. Verify all edits were applied successfully
2. Note the correction summary (e.g., "Updated 4 claims in section 2.1")
3. Remind user to commit changes
## Search best practices
### Query construction
**Good queries** (specific, current):
- "Claude Opus 4.5 context window 2026"
- "GPT-5.2 official release announcement"
- "Gemini 3 Pro token limit specifications"
**Poor queries** (vague, generic):
- "Claude context"
- "AI models"
- "Latest version"
### Source evaluation
**Prefer official sources:**
1. Product official pages (highest authority)
2. API documentation
3. Official blog announcements
4. GitHub releases (for open source)
**Use with caution:**
- Third-party aggregators (llm-stats.com, etc.) - verify against official sources
- Blog posts and articles - cross-reference claims
- Social media - only for announcements, verify elsewhere
**Avoid:**
- Outdated documentation
- Unofficial wikis without citations
- Speculation and rumors
### Handling ambiguity
When sources conflict:
1. Prioritize most recent official documentation
2. Note the discrepancy in the report
3. Present both sources to the user
4. Recommend contacting vendor if critical
When no source found:
1. Mark as ❓ Unverifiable
2. Suggest alternative phrasing: "According to [Source] as of [Date]..."
3. Recommend adding qualification: "approximately", "reported as"
## Special considerations
### Time-sensitive information
Always include temporal context:
**Good corrections:**
- "截至 2026 年 1 月" (As of January 2026)
- "Claude Sonnet 4.5 (released September 2025)"
**Poor corrections:**
- "Latest version" (becomes outdated)
- "Current model" (ambiguous timeframe)
### Numerical precision
Match precision to source:
**Source says:** "approximately 1 million tokens"
**Write:** "1M tokens (approximately)"
**Source says:** "200,000 token context window"
**Write:** "200K tokens" (exact)
### Citation format
Include citations in corrections:
```markdown
> **注**:具体上下文窗口以模型官方文档为准,本书写作时使用 Claude Sonnet 4.5 为主要工具。
```
Link to sources when possible.
## Examples
### Example 1: Technical specification update
**User request:** "Fact-check the AI model context windows in section 2.1"
**Process:**
1. Identify claims: Claude 3.5 Sonnet (200K), GPT-4o (128K), Gemini 1.5 Pro (2M)
2. Search official docs for current models
3. Find: Claude Sonnet 4.5, GPT-5.2, Gemini 3 Pro
4. Generate report showing discrepancies
5. Apply corrections after approval
### Example 2: Statistical data verification
**User request:** "Verify the benchmark scores in chapter 5"
**Process:**
1. Extract numerical claims
2. Search for official benchmark publications
3. Compare reported vs. source values
4. Flag any discrepancies with source links
5. Update with verified figures
### Example 3: Version number validation
**User request:** "Check if these library versions are still current"
**Process:**
1. List all version numbers mentioned
2. Check package registries (npm, PyPI, etc.)
3. Identify outdated versions
4. Suggest updates with changelog references
5. Update after user confirms
## Quality checklist
Before completing fact-check:
- [ ] All factual claims identified and categorized
- [ ] Each claim verified against official sources
- [ ] Sources are authoritative and current
- [ ] Correction report is clear and actionable
- [ ] Temporal context included where relevant
- [ ] User approval obtained before changes
- [ ] All edits verified successful
- [ ] Summary provided to user
## Limitations
**This skill cannot:**
- Verify subjective opinions or judgments
- Access paywalled or restricted sources
- Determine "truth" in disputed claims
- Predict future specifications or features
**For such cases:**
- Note the limitation in the report
- Suggest qualification language
- Recommend user research or expert consultation