docs: update all documentation for 17 source types

Update 32 documentation files across English and Chinese (zh-CN) docs
to reflect the 10 new source types added in the previous commit.

Updated files:
- README.md, README.zh-CN.md — taglines, feature lists, examples, install extras
- docs/reference/ — CLI_REFERENCE, FEATURE_MATRIX, MCP_REFERENCE, CONFIG_FORMAT, API_REFERENCE
- docs/features/ — UNIFIED_SCRAPING with generic merge docs
- docs/advanced/ — multi-source guide, MCP server guide
- docs/getting-started/ — installation extras, quick-start examples
- docs/user-guide/ — core-concepts, scraping, packaging, workflows (complex-merge)
- docs/ — FAQ, TROUBLESHOOTING, BEST_PRACTICES, ARCHITECTURE, UNIFIED_PARSERS, README
- Root — BULLETPROOF_QUICKSTART, CONTRIBUTING, ROADMAP
- docs/zh-CN/ — Chinese translations for all of the above

32 files changed, +3,016 lines, -245 lines
This commit is contained in:
yusyus
2026-03-15 15:56:04 +03:00
parent 53b911b697
commit 37cb307455
32 changed files with 3011 additions and 240 deletions

View File

@@ -1,6 +1,6 @@
# MCP Server Setup Guide
> **Skill Seekers v3.1.0**
> **Skill Seekers v3.2.0**
> **Integrate with AI agents via Model Context Protocol**
---
@@ -143,7 +143,7 @@ skill-seekers-mcp --transport http --port 8765
## Available Tools
26 tools organized by category:
27 tools organized by category:
### Core Tools (9)
- `list_configs` - List presets
@@ -156,9 +156,10 @@ skill-seekers-mcp --transport http --port 8765
- `enhance_skill` - AI enhancement
- `install_skill` - Complete workflow
### Extended Tools (9)
### Extended Tools (10)
- `scrape_github` - GitHub repo
- `scrape_pdf` - PDF extraction
- `scrape_generic` - Generic scraper for 10 new source types (see below)
- `scrape_codebase` - Local code
- `unified_scrape` - Multi-source
- `detect_patterns` - Pattern detection
@@ -180,6 +181,37 @@ skill-seekers-mcp --transport http --port 8765
- `export_to_faiss`
- `export_to_qdrant`
### scrape_generic Tool
The `scrape_generic` tool is the generic entry point for 10 new source types added in v3.2.0. It delegates to the appropriate CLI scraper module.
**Supported source types:** `jupyter`, `html`, `openapi`, `asciidoc`, `pptx`, `rss`, `manpage`, `confluence`, `notion`, `chat`
**Parameters:**
| Name | Type | Required | Description |
|------|------|----------|-------------|
| `source_type` | string | Yes | One of the 10 supported source types |
| `name` | string | Yes | Skill name for the output |
| `path` | string | No | File or directory path (for file-based sources) |
| `url` | string | No | URL (for URL-based sources like confluence, notion, rss) |
**Usage examples:**
```
"Scrape the Jupyter notebook analysis.ipynb"
→ scrape_generic(source_type="jupyter", name="analysis", path="analysis.ipynb")
"Extract content from the API spec"
→ scrape_generic(source_type="openapi", name="my-api", path="api-spec.yaml")
"Process the PowerPoint slides"
→ scrape_generic(source_type="pptx", name="slides", path="presentation.pptx")
"Scrape the Confluence wiki"
→ scrape_generic(source_type="confluence", name="wiki", url="https://wiki.example.com")
```
See [MCP Reference](../reference/MCP_REFERENCE.md) for full details.
---