Files
skill-seekers-reference/docs/reference/CLI_REFERENCE.md
yusyus 37cb307455 docs: update all documentation for 17 source types
Update 32 documentation files across English and Chinese (zh-CN) docs
to reflect the 10 new source types added in the previous commit.

Updated files:
- README.md, README.zh-CN.md — taglines, feature lists, examples, install extras
- docs/reference/ — CLI_REFERENCE, FEATURE_MATRIX, MCP_REFERENCE, CONFIG_FORMAT, API_REFERENCE
- docs/features/ — UNIFIED_SCRAPING with generic merge docs
- docs/advanced/ — multi-source guide, MCP server guide
- docs/getting-started/ — installation extras, quick-start examples
- docs/user-guide/ — core-concepts, scraping, packaging, workflows (complex-merge)
- docs/ — FAQ, TROUBLESHOOTING, BEST_PRACTICES, ARCHITECTURE, UNIFIED_PARSERS, README
- Root — BULLETPROOF_QUICKSTART, CONTRIBUTING, ROADMAP
- docs/zh-CN/ — Chinese translations for all of the above

32 files changed, +3,016 lines, -245 lines
2026-03-15 15:56:04 +03:00

1631 lines
40 KiB
Markdown

# CLI Reference - Skill Seekers
> **Version:** 3.2.0
> **Last Updated:** 2026-03-15
> **Complete reference for all 30 CLI commands**
---
## Table of Contents
- [Overview](#overview)
- [Installation](#installation)
- [Global Flags](#global-flags)
- [Environment Variables](#environment-variables)
- [Command Reference](#command-reference)
- [analyze](#analyze) - Analyze local codebase
- [asciidoc](#asciidoc) - Extract from AsciiDoc files
- [chat](#chat) - Extract from Slack/Discord
- [config](#config) - Configuration wizard
- [confluence](#confluence) - Extract from Confluence
- [create](#create) - Create skill (auto-detects source)
- [enhance](#enhance) - AI enhancement (local mode)
- [enhance-status](#enhance-status) - Monitor enhancement
- [estimate](#estimate) - Estimate page counts
- [github](#github) - Scrape GitHub repository
- [html](#html) - Extract from local HTML files
- [install](#install) - One-command complete workflow
- [install-agent](#install-agent) - Install to AI agent
- [jupyter](#jupyter) - Extract from Jupyter notebooks
- [manpage](#manpage) - Extract from man pages
- [multilang](#multilang) - Multi-language docs
- [notion](#notion) - Extract from Notion
- [openapi](#openapi) - Extract from OpenAPI/Swagger specs
- [package](#package) - Package skill for platform
- [pdf](#pdf) - Extract from PDF
- [pptx](#pptx) - Extract from PowerPoint files
- [quality](#quality) - Quality scoring
- [resume](#resume) - Resume interrupted jobs
- [rss](#rss) - Extract from RSS/Atom feeds
- [scrape](#scrape) - Scrape documentation
- [stream](#stream) - Stream large files
- [unified](#unified) - Multi-source scraping
- [update](#update) - Incremental updates
- [upload](#upload) - Upload to platform
- [video](#video) - Video extraction & setup
- [workflows](#workflows) - Manage workflow presets
- [Common Workflows](#common-workflows)
- [Exit Codes](#exit-codes)
- [Troubleshooting](#troubleshooting)
---
## Overview
Skill Seekers provides a unified CLI for converting documentation, GitHub repositories, PDFs, videos, notebooks, wikis, and 17 total source types into AI-ready skills for 16+ LLM platforms and RAG pipelines.
### Installation
```bash
# Basic installation
pip install skill-seekers
# With all platform support
pip install skill-seekers[all-llms]
# Development setup
pip install -e ".[all-llms,dev]"
```
Verify installation:
```bash
skill-seekers --version
```
### Global Flags
These flags work with **all scraper commands** (scrape, github, analyze, pdf, create):
| Flag | Description |
|------|-------------|
| `-h, --help` | Show help message and exit |
| `--version` | Show version number and exit |
| `-n, --name` | Skill name |
| `-d, --description` | Skill description |
| `-o, --output` | Output directory |
| `--enhance-level` | AI enhancement level (0-3) |
| `--api-key` | Anthropic API key |
| `-v, --verbose` | Enable verbose (DEBUG) output |
| `-q, --quiet` | Minimize output (WARNING only) |
| `--dry-run` | Preview without executing |
| `--enhance-workflow` | Apply enhancement workflow preset |
### Environment Variables
See [ENVIRONMENT_VARIABLES.md](ENVIRONMENT_VARIABLES.md) for complete reference.
**Common variables:**
| Variable | Purpose |
|----------|---------|
| `ANTHROPIC_API_KEY` | Claude AI API access |
| `GOOGLE_API_KEY` | Google Gemini API access |
| `OPENAI_API_KEY` | OpenAI API access |
| `GITHUB_TOKEN` | GitHub API (higher rate limits) |
---
## Command Reference
Commands are organized alphabetically.
---
### analyze
Analyze local codebase and extract code knowledge.
**Purpose:** Deep code analysis with pattern detection, API extraction, and documentation generation.
**Syntax:**
```bash
skill-seekers analyze --directory DIR [options]
```
**Arguments:**
| Name | Required | Description |
|------|----------|-------------|
| `--directory DIR` | Yes | Directory to analyze |
| `--output DIR` | No | Output directory (default: output/codebase/) |
**Flags:**
| Short | Long | Default | Description |
|-------|------|---------|-------------|
| `-n` | `--name` | auto | Skill name (defaults to directory name) |
| `-d` | `--description` | auto | Skill description |
| | `--preset` | standard | Analysis preset: quick, standard, comprehensive |
| | `--preset-list` | | Show available presets and exit |
| | `--languages` | auto | Comma-separated languages (Python,JavaScript,C++) |
| | `--file-patterns` | | Comma-separated file patterns |
| | `--enhance-level` | 0 | AI enhancement: 0=off (default), 1=SKILL.md, 2=+config, 3=full |
| | `--api-key` | | Anthropic API key (or ANTHROPIC_API_KEY env) |
| | `--enhance-workflow` | | Apply workflow preset (can use multiple) |
| | `--enhance-stage` | | Add inline enhancement stage (name:prompt) |
| | `--var` | | Override workflow variable (key=value) |
| | `--workflow-dry-run` | | Preview workflow without executing |
| | `--dry-run` | | Preview analysis without creating output |
| `-v` | `--verbose` | | Enable verbose (DEBUG) logging |
| `-q` | `--quiet` | | Minimize output (WARNING only) |
| | `--skip-api-reference` | | Skip API docs generation |
| | `--skip-dependency-graph` | | Skip dependency graph |
| | `--skip-patterns` | | Skip pattern detection |
| | `--skip-test-examples` | | Skip test example extraction |
| | `--skip-how-to-guides` | | Skip how-to guide generation |
| | `--skip-config-patterns` | | Skip config pattern extraction |
| | `--skip-docs` | | Skip project docs (README) |
| | `--no-comments` | | Skip comment extraction |
**Examples:**
```bash
# Basic analysis with defaults
skill-seekers analyze --directory ./my-project
# Quick analysis (1-2 min)
skill-seekers analyze --directory ./my-project --preset quick
# Comprehensive analysis with all features
skill-seekers analyze --directory ./my-project --preset comprehensive
# Specific languages only
skill-seekers analyze --directory ./my-project --languages Python,JavaScript
# Skip heavy features for faster analysis
skill-seekers analyze --directory ./my-project --skip-dependency-graph --skip-patterns
```
**Exit Codes:**
- `0` - Success
- `1` - Analysis failed
---
### asciidoc
Extract content from AsciiDoc files and generate skill.
**Purpose:** Convert `.adoc` / `.asciidoc` documentation into AI-ready skills.
**Syntax:**
```bash
skill-seekers asciidoc [options]
```
**Key Flags:**
| Flag | Description |
|------|-------------|
| `--asciidoc-path PATH` | Path to AsciiDoc file or directory |
| `-n, --name` | Skill name |
| `--from-json FILE` | Build from extracted JSON |
| `--enhance-level` | AI enhancement (default: 0) |
| `--dry-run` | Preview without executing |
**Examples:**
```bash
# Single file
skill-seekers asciidoc --asciidoc-path guide.adoc --name my-guide
# Directory of AsciiDoc files
skill-seekers asciidoc --asciidoc-path ./docs/ --name project-docs
```
---
### chat
Extract knowledge from Slack or Discord chat exports.
**Purpose:** Convert chat history into searchable AI-ready skills.
**Syntax:**
```bash
skill-seekers chat [options]
```
**Key Flags:**
| Flag | Description |
|------|-------------|
| `--export-path PATH` | Path to chat export directory or file |
| `--platform {slack,discord}` | Chat platform (default: slack) |
| `--token TOKEN` | API token for authentication |
| `--channel CHANNEL` | Channel name or ID to extract from |
| `--max-messages N` | Max messages to extract (default: 10000) |
| `-n, --name` | Skill name |
| `--dry-run` | Preview without executing |
**Examples:**
```bash
# From Slack export
skill-seekers chat --export-path ./slack-export/ --name team-knowledge
# From Discord via API
skill-seekers chat --platform discord --token $DISCORD_TOKEN --channel general --name discord-docs
```
---
### config
Interactive configuration wizard for API keys and settings.
**Purpose:** Setup GitHub tokens, API keys, and preferences.
**Syntax:**
```bash
skill-seekers config [options]
```
**Flags:**
| Short | Long | Description |
|-------|------|-------------|
| | `--github` | Go directly to GitHub token setup |
| | `--api-keys` | Go directly to API keys setup |
| | `--show` | Show current configuration |
| | `--test` | Test connections |
**Examples:**
```bash
# Full configuration wizard
skill-seekers config
# Quick GitHub setup
skill-seekers config --github
# View current config
skill-seekers config --show
# Test all connections
skill-seekers config --test
```
---
### confluence
Extract content from Confluence wikis.
**Purpose:** Convert Confluence spaces into AI-ready skills via API or HTML export.
**Syntax:**
```bash
skill-seekers confluence [options]
```
**Key Flags:**
| Flag | Description |
|------|-------------|
| `--base-url URL` | Confluence instance base URL |
| `--space-key KEY` | Confluence space key |
| `--export-path PATH` | Path to Confluence HTML/XML export directory |
| `--username USER` | Confluence username |
| `--token TOKEN` | Confluence API token |
| `--max-pages N` | Max pages to extract (default: 500) |
| `-n, --name` | Skill name |
| `--dry-run` | Preview without executing |
**Examples:**
```bash
# Via API
skill-seekers confluence --base-url https://wiki.example.com --space-key DEV \
--username user@example.com --token $CONFLUENCE_TOKEN --name dev-wiki
# From export
skill-seekers confluence --export-path ./confluence-export/ --name team-docs
```
---
### create
Create skill from any source. Auto-detects source type.
**Purpose:** Universal entry point - handles URLs, GitHub repos, local directories, PDFs, and config files automatically.
**Syntax:**
```bash
skill-seekers create [source] [options]
```
**Arguments:**
| Name | Required | Description |
|------|----------|-------------|
| `source` | No | Source URL, repo, path, or config file |
**Source Types (Auto-Detected):**
| Source Pattern | Type | Example |
|----------------|------|---------|
| `https://...` | Documentation | `https://docs.react.dev/` |
| `owner/repo` | GitHub | `facebook/react` |
| `./path` | Local codebase | `./my-project` |
| `*.pdf` | PDF | `manual.pdf` |
| `*.docx` | Word | `report.docx` |
| `*.epub` | EPUB | `book.epub` |
| `*.ipynb` | Jupyter Notebook | `analysis.ipynb` |
| `*.html`/`*.htm` | Local HTML | `docs.html` |
| `*.yaml`/`*.yml` | OpenAPI/Swagger | `openapi.yaml` |
| `*.adoc`/`*.asciidoc` | AsciiDoc | `guide.adoc` |
| `*.pptx` | PowerPoint | `slides.pptx` |
| `*.rss`/`*.atom` | RSS/Atom feed | `feed.rss` |
| `*.1`-`*.8`/`*.man` | Man page | `grep.1` |
| `*.json` | Config file | `config.json` |
**Flags:**
| Short | Long | Default | Description |
|-------|------|---------|-------------|
| `-n` | `--name` | auto | Skill name |
| `-d` | `--description` | auto | Skill description |
| `-o` | `--output` | auto | Output directory |
| `-p` | `--preset` | | Analysis preset: quick, standard, comprehensive |
| `-c` | `--config` | | Load settings from JSON file |
| | `--enhance-level` | 2 | AI enhancement level (0-3) |
| | `--api-key` | | Anthropic API key |
| | `--enhance-workflow` | | Apply workflow preset (can use multiple) |
| | `--enhance-stage` | | Add inline enhancement stage |
| | `--var` | | Override workflow variable (key=value) |
| | `--workflow-dry-run` | | Preview workflow without executing |
| | `--dry-run` | | Preview without creating |
| | `--chunk-for-rag` | | Enable RAG chunking |
| | `--chunk-tokens` | 512 | Chunk size in tokens |
| | `--chunk-overlap-tokens` | 50 | Chunk overlap in tokens |
| | `--help-web` | | Show web scraping options |
| | `--help-github` | | Show GitHub options |
| | `--help-local` | | Show local analysis options |
| | `--help-pdf` | | Show PDF options |
| | `--help-all` | | Show all 120+ options |
**Examples:**
```bash
# Documentation website
skill-seekers create https://docs.django.com/
# GitHub repository
skill-seekers create facebook/react
# Local codebase
skill-seekers create ./my-project
# PDF file
skill-seekers create manual.pdf --name product-docs
# With preset
skill-seekers create https://docs.react.dev/ --preset quick
# With enhancement workflow
skill-seekers create ./my-project --enhance-workflow security-focus
# Multi-workflow chaining
skill-seekers create ./my-project \
--enhance-workflow security-focus \
--enhance-workflow api-documentation
```
---
### enhance
Enhance SKILL.md using local coding agent (Claude Code).
**Purpose:** AI-powered quality improvement without API costs. Requires Claude Code installed.
**Syntax:**
```bash
skill-seekers enhance SKILL_DIRECTORY [options]
```
**Arguments:**
| Name | Required | Description |
|------|----------|-------------|
| `SKILL_DIRECTORY` | Yes | Path to skill directory |
**Flags:**
| Short | Long | Default | Description |
|-------|------|---------|-------------|
| | `--agent` | claude | Local coding agent to use |
| | `--agent-cmd` | | Override agent command template |
| | `--background` | | Run in background |
| | `--daemon` | | Run as daemon |
| | `--no-force` | | Enable confirmations |
| | `--timeout` | 600 | Timeout in seconds |
**Examples:**
```bash
# Basic enhancement
skill-seekers enhance output/react/
# Background mode
skill-seekers enhance output/react/ --background
# With custom timeout
skill-seekers enhance output/react/ --timeout 1200
# Monitor background enhancement
skill-seekers enhance-status output/react/ --watch
```
**Requirements:** Claude Code must be installed and authenticated.
---
### enhance-status
Monitor background enhancement processes.
**Purpose:** Check status of enhancement running in background/daemon mode.
**Syntax:**
```bash
skill-seekers enhance-status SKILL_DIRECTORY [options]
```
**Arguments:**
| Name | Required | Description |
|------|----------|-------------|
| `SKILL_DIRECTORY` | Yes | Path to skill directory |
**Flags:**
| Short | Long | Default | Description |
|-------|------|---------|-------------|
| `-w` | `--watch` | | Watch in real-time |
| | `--json` | | JSON output |
| | `--interval` | 5 | Watch interval in seconds |
**Examples:**
```bash
# Check status once
skill-seekers enhance-status output/react/
# Watch continuously
skill-seekers enhance-status output/react/ --watch
# JSON output for scripting
skill-seekers enhance-status output/react/ --json
```
---
### estimate
Estimate page count before scraping.
**Purpose:** Preview how many pages will be scraped without downloading.
**Syntax:**
```bash
skill-seekers estimate [config] [options]
```
**Arguments:**
| Name | Required | Description |
|------|----------|-------------|
| `config` | No | Config JSON file path |
**Flags:**
| Short | Long | Default | Description |
|-------|------|---------|-------------|
| | `--all` | | List all available configs |
| | `--max-discovery` | 1000 | Max pages to discover |
**Examples:**
```bash
# Estimate with config file
skill-seekers estimate configs/react.json
# Quick estimate (100 pages)
skill-seekers estimate configs/react.json --max-discovery 100
# List all available presets
skill-seekers estimate --all
```
---
### github
Scrape GitHub repository and generate skill.
**Purpose:** Extract code, issues, releases, and metadata from GitHub repos.
**Syntax:**
```bash
skill-seekers github [options]
```
**Flags:**
| Short | Long | Default | Description |
|-------|------|---------|-------------|
| | `--repo` | | Repository (owner/repo format) |
| `-c` | `--config` | | Config JSON file |
| | `--token` | | GitHub personal access token |
| `-n` | `--name` | auto | Skill name |
| `-d` | `--description` | auto | Description |
| `-o` | `--output` | auto | Output directory |
| | `--no-issues` | | Skip GitHub issues |
| | `--no-changelog` | | Skip CHANGELOG |
| | `--no-releases` | | Skip releases |
| | `--max-issues` | 100 | Max issues to fetch |
| | `--scrape-only` | | Only scrape, don't build |
| | `--enhance-level` | 2 | AI enhancement (0-3) |
| | `--api-key` | | Anthropic API key |
| | `--enhance-workflow` | | Apply workflow preset |
| | `--non-interactive` | | CI/CD mode (fail fast) |
| | `--profile` | | GitHub profile from config |
| | `--dry-run` | | Preview without executing |
| `-v` | `--verbose` | | Enable verbose (DEBUG) logging |
| `-q` | `--quiet` | | Minimize output (WARNING only) |
**Examples:**
```bash
# Basic repo analysis
skill-seekers github --repo facebook/react
# With GitHub token (higher rate limits)
skill-seekers github --repo facebook/react --token $GITHUB_TOKEN
# Skip issues for faster scraping
skill-seekers github --repo facebook/react --no-issues
# Dry run to preview
skill-seekers github --repo facebook/react --dry-run
# Scrape only, build later
skill-seekers github --repo facebook/react --scrape-only
```
---
### html
Extract content from local HTML files and generate skill.
**Purpose:** Convert local HTML documentation into AI-ready skills (for offline/exported docs).
**Syntax:**
```bash
skill-seekers html [options]
```
**Key Flags:**
| Flag | Description |
|------|-------------|
| `--html-path PATH` | Path to HTML file or directory |
| `-n, --name` | Skill name |
| `--from-json FILE` | Build from extracted JSON |
| `--enhance-level` | AI enhancement (default: 0) |
| `--dry-run` | Preview without executing |
**Examples:**
```bash
# Single HTML file
skill-seekers html --html-path docs/index.html --name my-docs
# Directory of HTML files
skill-seekers html --html-path ./html-export/ --name exported-docs
```
---
### install
One-command complete workflow: fetch → scrape → enhance → package → upload.
**Purpose:** End-to-end automation for common workflows.
**Syntax:**
```bash
skill-seekers install --config CONFIG [options]
```
**Arguments:**
| Name | Required | Description |
|------|----------|-------------|
| `--config CONFIG` | Yes | Config name or path |
**Flags:**
| Short | Long | Default | Description |
|-------|------|---------|-------------|
| | `--destination` | output/ | Output directory |
| | `--no-upload` | | Skip upload to Claude |
| | `--unlimited` | | Remove page limits |
| | `--dry-run` | | Preview without executing |
**Examples:**
```bash
# Complete workflow with preset
skill-seekers install --config react
# Skip upload
skill-seekers install --config react --no-upload
# Custom config
skill-seekers install --config configs/my-project.json
# Dry run to preview
skill-seekers install --config react --dry-run
```
**Note:** AI enhancement is mandatory for install command.
---
### install-agent
Install skill to AI agent directories (Cursor, Windsurf, Cline).
**Purpose:** Direct installation to IDE AI assistant context directories.
**Syntax:**
```bash
skill-seekers install-agent SKILL_DIRECTORY --agent AGENT [options]
```
**Arguments:**
| Name | Required | Description |
|------|----------|-------------|
| `SKILL_DIRECTORY` | Yes | Path to skill directory |
| `--agent AGENT` | Yes | Target agent: cursor, windsurf, cline, continue |
**Flags:**
| Short | Long | Description |
|-------|------|-------------|
| | `--force` | Overwrite existing |
**Examples:**
```bash
# Install to Cursor
skill-seekers install-agent output/react/ --agent cursor
# Install to Windsurf
skill-seekers install-agent output/react/ --agent windsurf
# Force overwrite
skill-seekers install-agent output/react/ --agent cursor --force
```
---
### jupyter
Extract content from Jupyter Notebook files and generate skill.
**Purpose:** Convert `.ipynb` notebooks into AI-ready skills with code, markdown, and outputs.
**Syntax:**
```bash
skill-seekers jupyter [options]
```
**Key Flags:**
| Flag | Description |
|------|-------------|
| `--notebook PATH` | Path to .ipynb file or directory |
| `-n, --name` | Skill name |
| `--from-json FILE` | Build from extracted JSON |
| `--enhance-level` | AI enhancement (default: 0) |
| `--dry-run` | Preview without executing |
**Examples:**
```bash
# Single notebook
skill-seekers jupyter --notebook analysis.ipynb --name data-analysis
# Directory of notebooks
skill-seekers jupyter --notebook ./notebooks/ --name ml-tutorials
```
---
### manpage
Extract content from Unix/Linux man pages and generate skill.
**Purpose:** Convert man pages into AI-ready reference skills.
**Syntax:**
```bash
skill-seekers manpage [options]
```
**Key Flags:**
| Flag | Description |
|------|-------------|
| `--man-names NAMES` | Comma-separated man page names (e.g., `ls,grep,find`) |
| `--man-path PATH` | Path to directory containing man page files |
| `--sections SECTIONS` | Comma-separated section numbers (e.g., `1,3,8`) |
| `-n, --name` | Skill name |
| `--dry-run` | Preview without executing |
**Examples:**
```bash
# By name (system man pages)
skill-seekers manpage --man-names ls,grep,find,awk --name unix-essentials
# From directory
skill-seekers manpage --man-path /usr/share/man/man1/ --sections 1 --name section1-cmds
```
---
### multilang
Multi-language documentation support.
**Purpose:** Scrape and merge documentation in multiple languages.
**Syntax:**
```bash
skill-seekers multilang --config CONFIG [options]
```
**Flags:**
| Short | Long | Description |
|-------|------|-------------|
| `-c` | `--config` | Config JSON file |
| | `--primary` | Primary language |
| | `--languages` | Comma-separated languages |
| | `--merge-strategy` | How to merge: parallel, hierarchical |
**Examples:**
```bash
# Multi-language scrape
skill-seekers multilang --config configs/react-i18n.json
# Specific languages
skill-seekers multilang --config configs/docs.json --languages en,zh,es
```
---
### notion
Extract content from Notion workspaces.
**Purpose:** Convert Notion pages and databases into AI-ready skills via API or export.
**Syntax:**
```bash
skill-seekers notion [options]
```
**Key Flags:**
| Flag | Description |
|------|-------------|
| `--database-id ID` | Notion database ID to extract from |
| `--page-id ID` | Notion page ID to extract from |
| `--export-path PATH` | Path to Notion export directory |
| `--token TOKEN` | Notion integration token |
| `--max-pages N` | Max pages to extract (default: 500) |
| `-n, --name` | Skill name |
| `--dry-run` | Preview without executing |
**Examples:**
```bash
# Via API
skill-seekers notion --database-id abc123 --token $NOTION_TOKEN --name team-docs
# From export
skill-seekers notion --export-path ./notion-export/ --name project-wiki
```
---
### openapi
Extract content from OpenAPI/Swagger specifications and generate skill.
**Purpose:** Convert API specs into AI-ready reference skills with endpoint documentation.
**Syntax:**
```bash
skill-seekers openapi [options]
```
**Key Flags:**
| Flag | Description |
|------|-------------|
| `--spec PATH` | Path to OpenAPI/Swagger spec file |
| `--spec-url URL` | URL to OpenAPI/Swagger spec |
| `-n, --name` | Skill name |
| `--from-json FILE` | Build from extracted JSON |
| `--enhance-level` | AI enhancement (default: 0) |
| `--dry-run` | Preview without executing |
**Examples:**
```bash
# From local file
skill-seekers openapi --spec api/openapi.yaml --name my-api
# From URL
skill-seekers openapi --spec-url https://petstore.swagger.io/v2/swagger.json --name petstore
```
---
### package
Package skill directory into platform-specific format.
**Purpose:** Create uploadable packages for Claude, Gemini, OpenAI, and RAG platforms.
**Syntax:**
```bash
skill-seekers package SKILL_DIRECTORY [options]
```
**Arguments:**
| Name | Required | Description |
|------|----------|-------------|
| `SKILL_DIRECTORY` | Yes | Path to skill directory |
**Flags:**
| Short | Long | Default | Description |
|-------|------|---------|-------------|
| | `--target` | claude | Target platform |
| | `--no-open` | | Don't open output folder |
| | `--skip-quality-check` | | Skip quality checks |
| | `--upload` | | Auto-upload after packaging |
| | `--streaming` | | Streaming mode for large docs |
| | `--streaming-chunk-chars` | 4000 | Max chars per chunk (streaming) |
| | `--streaming-overlap-chars` | 200 | Overlap between chunks (chars) |
| | `--batch-size` | 100 | Chunks per batch |
| | `--chunk-for-rag` | | Enable RAG chunking |
| | `--chunk-tokens` | 512 | Max tokens per chunk |
| | `--chunk-overlap-tokens` | 50 | Overlap between chunks (tokens) |
| | `--no-preserve-code-blocks` | | Allow code block splitting |
**Supported Platforms:**
| Platform | Format | Flag |
|----------|--------|------|
| Claude AI | ZIP + YAML | `--target claude` |
| Google Gemini | tar.gz | `--target gemini` |
| OpenAI | ZIP + Vector | `--target openai` |
| LangChain | Documents | `--target langchain` |
| LlamaIndex | TextNodes | `--target llama-index` |
| Haystack | Documents | `--target haystack` |
| ChromaDB | Collection | `--target chroma` |
| Weaviate | Objects | `--target weaviate` |
| Qdrant | Points | `--target qdrant` |
| FAISS | Index | `--target faiss` |
| Pinecone | Markdown | `--target pinecone` |
| Markdown | ZIP | `--target markdown` |
**Examples:**
```bash
# Package for Claude (default)
skill-seekers package output/react/
# Package for Gemini
skill-seekers package output/react/ --target gemini
# Package for multiple platforms
for platform in claude gemini openai; do
skill-seekers package output/react/ --target $platform
done
# Package with upload
skill-seekers package output/react/ --target claude --upload
# Streaming mode for large docs
skill-seekers package output/large-docs/ --streaming
```
---
### pdf
Extract content from PDF and generate skill.
**Purpose:** Convert PDF manuals, documentation, and papers into skills.
**Syntax:**
```bash
skill-seekers pdf [options]
```
**Flags:**
| Short | Long | Default | Description |
|-------|------|---------|-------------|
| `-c` | `--config` | | PDF config JSON file |
| | `--pdf` | | Direct PDF file path |
| `-n` | `--name` | auto | Skill name |
| `-d` | `--description` | auto | Description |
| `-o` | `--output` | auto | Output directory |
| | `--from-json` | | Build from extracted JSON |
| | `--enhance-level` | 0 | AI enhancement (default: 0 for PDF) |
| | `--api-key` | | Anthropic API key |
| | `--enhance-workflow` | | Apply workflow preset |
| | `--enhance-stage` | | Add inline stage |
| | `--var` | | Override workflow variable |
| | `--workflow-dry-run` | | Preview workflow |
| | `--dry-run` | | Preview without executing |
| `-v` | `--verbose` | | Enable verbose (DEBUG) logging |
| `-q` | `--quiet` | | Minimize output (WARNING only) |
**Examples:**
```bash
# Direct PDF path
skill-seekers pdf --pdf manual.pdf --name product-manual
# With config file
skill-seekers pdf --config configs/manual.json
# Enable enhancement
skill-seekers pdf --pdf manual.pdf --enhance-level 2
# Dry run to preview
skill-seekers pdf --pdf manual.pdf --name test --dry-run
```
---
### pptx
Extract content from PowerPoint files and generate skill.
**Purpose:** Convert `.pptx` presentations into AI-ready skills.
**Syntax:**
```bash
skill-seekers pptx [options]
```
**Key Flags:**
| Flag | Description |
|------|-------------|
| `--pptx PATH` | Path to PowerPoint file (.pptx) |
| `-n, --name` | Skill name |
| `--from-json FILE` | Build from extracted JSON |
| `--enhance-level` | AI enhancement (default: 0) |
| `--dry-run` | Preview without executing |
**Examples:**
```bash
# Extract from presentation
skill-seekers pptx --pptx training-slides.pptx --name training-material
# With enhancement
skill-seekers pptx --pptx architecture.pptx --name arch-overview --enhance-level 2
```
---
### quality
Analyze and score skill documentation quality.
**Purpose:** Quality assurance before packaging/uploading.
**Syntax:**
```bash
skill-seekers quality SKILL_DIRECTORY [options]
```
**Arguments:**
| Name | Required | Description |
|------|----------|-------------|
| `SKILL_DIRECTORY` | Yes | Path to skill directory |
**Flags:**
| Short | Long | Description |
|-------|------|-------------|
| | `--report` | Generate detailed report |
| | `--threshold` | Quality threshold (0-10) |
**Examples:**
```bash
# Basic quality check
skill-seekers quality output/react/
# Detailed report
skill-seekers quality output/react/ --report
# Fail if below threshold
skill-seekers quality output/react/ --threshold 7.0
```
---
### resume
Resume interrupted scraping job from checkpoint.
**Purpose:** Continue from where a scrape failed or was interrupted.
**Syntax:**
```bash
skill-seekers resume [JOB_ID] [options]
```
**Arguments:**
| Name | Required | Description |
|------|----------|-------------|
| `JOB_ID` | No | Job ID to resume |
**Flags:**
| Short | Long | Description |
|-------|------|-------------|
| | `--list` | List all resumable jobs |
| | `--clean` | Clean up old progress files |
**Examples:**
```bash
# List resumable jobs
skill-seekers resume --list
# Resume specific job
skill-seekers resume job-abc123
# Clean old checkpoints
skill-seekers resume --clean
```
---
### rss
Extract content from RSS/Atom feeds and generate skill.
**Purpose:** Convert blog feeds and news sources into AI-ready skills.
**Syntax:**
```bash
skill-seekers rss [options]
```
**Key Flags:**
| Flag | Description |
|------|-------------|
| `--feed-url URL` | URL of the RSS/Atom feed |
| `--feed-path PATH` | Path to local RSS/Atom feed file |
| `--follow-links` | Follow article links for full content (default: true) |
| `--no-follow-links` | Use feed summary only |
| `--max-articles N` | Max articles to extract (default: 50) |
| `-n, --name` | Skill name |
| `--dry-run` | Preview without executing |
**Examples:**
```bash
# From URL
skill-seekers rss --feed-url https://blog.example.com/feed.xml --name blog-knowledge
# From local file, summaries only
skill-seekers rss --feed-path ./feed.rss --no-follow-links --name feed-summaries
```
---
### scrape
Scrape documentation website and generate skill.
**Purpose:** The main command for converting web documentation into skills.
**Syntax:**
```bash
skill-seekers scrape [url] [options]
```
**Arguments:**
| Name | Required | Description |
|------|----------|-------------|
| `url` | No | Base documentation URL |
**Flags:**
| Short | Long | Default | Description |
|-------|------|---------|-------------|
| `-c` | `--config` | | Config JSON file |
| `-n` | `--name` | | Skill name |
| `-d` | `--description` | | Description |
| | `--enhance-level` | 2 | AI enhancement (0-3) |
| | `--api-key` | | Anthropic API key |
| | `--enhance-workflow` | | Apply workflow preset |
| | `--enhance-stage` | | Add inline stage |
| | `--var` | | Override workflow variable |
| | `--workflow-dry-run` | | Preview workflow |
| `-i` | `--interactive` | | Interactive mode |
| | `--url` | | Base URL (alternative to positional) |
| | `--max-pages` | | Max pages to scrape |
| | `--skip-scrape` | | Use existing data |
| | `--dry-run` | | Preview without scraping |
| | `--resume` | | Resume from checkpoint |
| | `--fresh` | | Clear checkpoint |
| `-r` | `--rate-limit` | 0.5 | Rate limit in seconds |
| `-w` | `--workers` | 1 | Parallel workers (max 10) |
| | `--async` | | Enable async mode |
| | `--no-rate-limit` | | Disable rate limiting |
| | `--interactive-enhancement` | | Interactive enhancement |
| `-v` | `--verbose` | | Verbose output |
| `-q` | `--quiet` | | Quiet output |
**Examples:**
```bash
# With preset config
skill-seekers scrape --config configs/react.json
# Quick mode
skill-seekers scrape --name react --url https://react.dev/
# Interactive mode
skill-seekers scrape --interactive
# Dry run
skill-seekers scrape --config configs/react.json --dry-run
# Fast async scraping
skill-seekers scrape --config configs/react.json --async --workers 5
# Skip scrape, rebuild from cache
skill-seekers scrape --config configs/react.json --skip-scrape
# Resume interrupted scrape
skill-seekers scrape --config configs/react.json --resume
```
---
### stream
Stream large files chunk-by-chunk.
**Purpose:** Memory-efficient processing for very large documentation sites.
**Syntax:**
```bash
skill-seekers stream --config CONFIG [options]
```
**Flags:**
| Short | Long | Description |
|-------|------|-------------|
| `-c` | `--config` | Config JSON file |
| | `--streaming-chunk-chars` | Maximum characters per chunk (default: 4000) |
| | `--output` | Output directory |
**Examples:**
```bash
# Stream large documentation
skill-seekers stream --config configs/large-docs.json
# Custom chunk size
skill-seekers stream --config configs/large-docs.json --streaming-chunk-chars 1000
```
---
### unified
Multi-source scraping combining docs + GitHub + PDF.
**Purpose:** Create a single skill from multiple sources with conflict detection.
**Syntax:**
```bash
skill-seekers unified --config FILE [options]
```
**Arguments:**
| Name | Required | Description |
|------|----------|-------------|
| `--config FILE` | Yes | Unified config JSON file |
**Flags:**
| Short | Long | Default | Description |
|-------|------|---------|-------------|
| | `--merge-mode` | claude-enhanced | Merge mode: rule-based, claude-enhanced |
| | `--fresh` | | Clear existing data |
| | `--dry-run` | | Dry run mode |
| | `--enhance-level` | | Override enhancement level (0-3) |
| | `--api-key` | | Anthropic API key (or ANTHROPIC_API_KEY env) |
| | `--enhance-workflow` | | Apply workflow preset (can use multiple) |
| | `--enhance-stage` | | Add inline enhancement stage (name:prompt) |
| | `--var` | | Override workflow variable (key=value) |
| | `--workflow-dry-run` | | Preview workflow without executing |
| | `--skip-codebase-analysis` | | Skip C3.x codebase analysis for GitHub sources |
**Examples:**
```bash
# Unified scraping
skill-seekers unified --config configs/react-unified.json
# Fresh start
skill-seekers unified --config configs/react-unified.json --fresh
# Rule-based merging
skill-seekers unified --config configs/react-unified.json --merge-mode rule-based
```
**Config Format:**
```json
{
"name": "react-complete",
"sources": [
{"type": "docs", "base_url": "https://react.dev/"},
{"type": "github", "repo": "facebook/react"}
]
}
```
---
### update
Update docs without full rescrape.
**Purpose:** Incremental updates for changed documentation.
**Syntax:**
```bash
skill-seekers update --config CONFIG [options]
```
**Flags:**
| Short | Long | Description |
|-------|------|-------------|
| `-c` | `--config` | Config JSON file |
| | `--since` | Update since date |
| | `--check-only` | Check for updates only |
**Examples:**
```bash
# Check for updates
skill-seekers update --config configs/react.json --check-only
# Update since specific date
skill-seekers update --config configs/react.json --since 2026-01-01
# Full update
skill-seekers update --config configs/react.json
```
---
### upload
Upload skill package to LLM platform or vector database.
**Purpose:** Deploy packaged skills to target platforms.
**Syntax:**
```bash
skill-seekers upload PACKAGE_FILE [options]
```
**Arguments:**
| Name | Required | Description |
|------|----------|-------------|
| `PACKAGE_FILE` | Yes | Path to package file (.zip, .tar.gz) |
**Flags:**
| Short | Long | Default | Description |
|-------|------|---------|-------------|
| | `--target` | claude | Target platform |
| | `--api-key` | | Platform API key |
| | `--chroma-url` | | ChromaDB URL |
| | `--persist-directory` | ./chroma_db | ChromaDB local directory |
| | `--embedding-function` | | Embedding function |
| | `--openai-api-key` | | OpenAI key for embeddings |
| | `--weaviate-url` | | Weaviate URL |
| | `--use-cloud` | | Use Weaviate Cloud |
| | `--cluster-url` | | Weaviate Cloud cluster URL |
**Examples:**
```bash
# Upload to Claude
skill-seekers upload output/react-claude.zip
# Upload to Gemini
skill-seekers upload output/react-gemini.tar.gz --target gemini
# Upload to ChromaDB
skill-seekers upload output/react-chroma.zip --target chroma
# Upload to Weaviate Cloud
skill-seekers upload output/react-weaviate.zip --target weaviate \
--use-cloud --cluster-url https://xxx.weaviate.network
```
---
### video
Extract skills from video tutorials (YouTube, Vimeo, or local files).
### Usage
```bash
# Setup (first time — auto-detects GPU, installs PyTorch + visual deps)
skill-seekers video --setup
# Extract from YouTube
skill-seekers video --url https://www.youtube.com/watch?v=VIDEO_ID --name my-skill
# With visual frame extraction (requires --setup first)
skill-seekers video --url VIDEO_URL --name my-skill --visual
# Local video file
skill-seekers video --url /path/to/video.mp4 --name my-skill
```
### Key Flags
| Flag | Description |
|------|-------------|
| `--setup` | Auto-detect GPU and install visual extraction dependencies |
| `--url URL` | Video URL (YouTube, Vimeo) or local file path |
| `--name NAME` | Skill name for output |
| `--visual` | Enable visual frame extraction (OCR on keyframes) |
| `--vision-api` | Use Claude Vision API as OCR fallback for low-confidence frames |
### Notes
- `--setup` detects NVIDIA (CUDA), AMD (ROCm), or CPU-only and installs the correct PyTorch variant
- Requires `pip install skill-seekers[video]` (transcripts) or `skill-seekers[video-full]` (+ whisper + scene detection)
- EasyOCR is NOT included in pip extras — it is installed by `--setup` with the correct GPU backend
---
### workflows
Manage enhancement workflow presets.
**Purpose:** List, inspect, copy, add, remove, and validate YAML workflow presets.
**Syntax:**
```bash
skill-seekers workflows ACTION [options]
```
**Actions:**
| Action | Description |
|--------|-------------|
| `list` | List all workflows (bundled + user) |
| `show` | Print YAML content of workflow |
| `copy` | Copy bundled workflow to user dir |
| `add` | Install custom YAML workflow |
| `remove` | Delete user workflow |
| `validate` | Validate workflow file |
**Flags:**
| Short | Long | Description |
|-------|------|-------------|
| | `--name` | Custom name for add action |
**Examples:**
```bash
# List all workflows
skill-seekers workflows list
# Show workflow content
skill-seekers workflows show security-focus
# Copy for editing
skill-seekers workflows copy security-focus
# Add custom workflow
skill-seekers workflows add ./my-workflow.yaml
# Add with custom name
skill-seekers workflows add ./workflow.yaml --name my-custom
# Remove user workflow
skill-seekers workflows remove my-workflow
# Validate workflow
skill-seekers workflows validate security-focus
skill-seekers workflows validate ./my-workflow.yaml
```
**Built-in Presets:**
- `default` - Standard enhancement
- `minimal` - Light enhancement
- `security-focus` - Security analysis (4 stages)
- `architecture-comprehensive` - Deep architecture review (7 stages)
- `api-documentation` - API docs focus (3 stages)
---
## Common Workflows
### Workflow 1: Documentation → Skill
```bash
# 1. Estimate pages (optional)
skill-seekers estimate configs/react.json
# 2. Scrape documentation
skill-seekers scrape --config configs/react.json
# 3. Enhance SKILL.md (optional, recommended)
skill-seekers enhance output/react/
# 4. Package for Claude
skill-seekers package output/react/ --target claude
# 5. Upload
skill-seekers upload output/react-claude.zip
```
### Workflow 2: GitHub → Skill
```bash
# 1. Analyze repository
skill-seekers github --repo facebook/react
# 2. Package
skill-seekers package output/react/ --target claude
# 3. Upload
skill-seekers upload output/react-claude.zip
```
### Workflow 3: Local Codebase → Skill
```bash
# 1. Analyze codebase
skill-seekers analyze --directory ./my-project
# 2. Package
skill-seekers package output/codebase/ --target claude
# 3. Install to Cursor
skill-seekers install-agent output/codebase/ --agent cursor
```
### Workflow 4: PDF → Skill
```bash
# 1. Extract PDF
skill-seekers pdf --pdf manual.pdf --name product-docs
# 2. Package
skill-seekers package output/product-docs/ --target claude
```
### Workflow 5: Multi-Source → Skill
```bash
# 1. Create unified config (configs/my-project.json)
# 2. Run unified scraping
skill-seekers unified --config configs/my-project.json
# 3. Package
skill-seekers package output/my-project/ --target claude
```
### Workflow 6: One-Command Complete
```bash
# Everything in one command
skill-seekers install --config react --destination ./output
# Or with create
skill-seekers create https://docs.react.dev/ --preset standard
```
---
## Exit Codes
| Code | Meaning |
|------|---------|
| `0` | Success |
| `1` | General error |
| `2` | Warning (e.g., estimation hit limit) |
| `130` | Interrupted by user (Ctrl+C) |
---
## Troubleshooting
### Command not found
```bash
# Ensure package is installed
pip install skill-seekers
# Check PATH
which skill-seekers
```
### ImportError
```bash
# Install in editable mode (development)
pip install -e .
```
### Rate limiting
```bash
# Increase rate limit
skill-seekers scrape --config react.json --rate-limit 1.0
```
### Out of memory
```bash
# Use streaming mode
skill-seekers package output/large/ --streaming
```
---
## See Also
- [Config Format](CONFIG_FORMAT.md) - JSON configuration specification
- [Environment Variables](ENVIRONMENT_VARIABLES.md) - Complete env var reference
- [MCP Reference](MCP_REFERENCE.md) - MCP tools documentation
---
*For additional help: `skill-seekers --help` or `skill-seekers <command> --help`*