This commit includes two major improvements:
## 1. Unified Create Command (v3.0.0 feature)
- Auto-detects source type (web, GitHub, local, PDF, config)
- Three-tier argument organization (universal, source-specific, advanced)
- Routes to existing scrapers (100% backward compatible)
- Progressive disclosure: 15 universal flags in default help
**New files:**
- src/skill_seekers/cli/source_detector.py - Auto-detection logic
- src/skill_seekers/cli/arguments/create.py - Argument definitions
- src/skill_seekers/cli/create_command.py - Main orchestrator
- src/skill_seekers/cli/parsers/create_parser.py - Parser integration
**Tests:**
- tests/test_source_detector.py (35 tests)
- tests/test_create_arguments.py (30 tests)
- tests/test_create_integration_basic.py (10 tests)
## 2. Enhanced Flag Consolidation (Phase 1)
- Consolidated 3 flags (--enhance, --enhance-local, --enhance-level) → 1 flag
- --enhance-level 0-3 with auto-detection of API vs LOCAL mode
- Default: --enhance-level 2 (balanced enhancement)
**Modified files:**
- arguments/{common,create,scrape,github,analyze}.py - Added enhance_level
- {doc_scraper,github_scraper,config_extractor,main}.py - Updated logic
- create_command.py - Uses consolidated flag
**Auto-detection:**
- If ANTHROPIC_API_KEY set → API mode
- Else → LOCAL mode (Claude Code)
## 3. PresetManager Bug Fix
- Fixed module naming conflict (presets.py vs presets/ directory)
- Moved presets.py → presets/manager.py
- Updated __init__.py exports
**Test Results:**
- All 160+ tests passing
- Zero regressions
- 100% backward compatible
Co-Authored-By: Claude Sonnet 4.5 <noreply@anthropic.com>
13 KiB
🚀 Skill Seekers v3.0.0 - LAUNCH BLITZ (One Week)
Strategy: Concentrated all-channel launch over 5 days
Goal: Maximum impact through simultaneous multi-platform release
📊 WHAT WE HAVE (All Ready)
| Component | Status |
|---|---|
| Code | ✅ v3.0.0 tagged, all tests pass |
| PyPI | ✅ Ready to publish |
| Website | ✅ Blog live with 4 posts |
| Docs | ✅ 18 integration guides ready |
| Examples | ✅ 12 working examples |
🎯 THE BLITZ STRATEGY
Instead of spreading over 4 weeks, we hit ALL channels simultaneously over 5 days. This creates a "surge" effect - people see us everywhere at once.
📅 5-DAY LAUNCH TIMELINE
DAY 1: Foundation (Monday)
Theme: "Release Day"
Morning (9-11 AM EST - Optimal Time)
-
Publish to PyPI
python -m build python -m twine upload dist/* -
Create GitHub Release
- Title: "v3.0.0 - Universal Intelligence Platform"
- Copy CHANGELOG v3.0.0 section
- Add release assets (optional)
Afternoon (1-3 PM EST)
- Publish main blog post on website
- Title: "Skill Seekers v3.0.0: The Universal Intelligence Platform"
- Share on personal Twitter/LinkedIn
Evening (Check metrics, respond to comments)
DAY 2: Social Media Blast (Tuesday)
Theme: "Social Surge"
Morning (9-11 AM EST)
Twitter/X Thread (10 tweets)
Tweet 1: 🚀 Skill Seekers v3.0.0 is LIVE!
The universal documentation preprocessor for AI systems.
16 output formats. 1,852 tests. One tool for LangChain, LlamaIndex, Cursor, Claude, and more.
Thread 🧵
---
Tweet 2: The Problem
Every AI project needs documentation ingestion.
But everyone rebuilds the same scraper:
- Handle pagination
- Extract clean text
- Chunk properly
- Add metadata
- Format for their tool
Stop rebuilding. Start using.
---
Tweet 3: Meet Skill Seekers v3.0.0
One command → Any format
pip install skill-seekers
skill-seekers scrape --config react.json
Output options:
- LangChain Documents
- LlamaIndex Nodes
- Claude skills
- Cursor rules
- Markdown for any vector DB
---
Tweet 4: For RAG Pipelines
Before: 50 lines of custom scraping code
After: 1 command
skill-seekers scrape --format langchain --config docs.json
Returns structured Document objects with metadata.
Ready for Chroma, Pinecone, Weaviate.
---
Tweet 5: For AI Coding Tools
Give Cursor complete framework knowledge:
skill-seekers scrape --target claude --config react.json
cp output/react-claude/.cursorrules ./
Now Cursor knows React better than most devs.
Also works with: Windsurf, Cline, Continue.dev
---
Tweet 6: 26 MCP Tools
Your AI agent can now prepare its own knowledge:
- scrape_docs
- scrape_github
- scrape_pdf
- package_skill
- install_skill
- And 21 more...
Your AI agent can prep its own knowledge.
---
Tweet 7: 1,852 Tests
Production-ready means tested.
- 100 test files
- 1,852 test cases
- CI/CD on every commit
- Multi-platform validation
This isn't a prototype. It's infrastructure.
---
Tweet 8: Cloud & CI/CD
AWS S3, GCS, Azure support.
GitHub Action ready.
Docker image available.
skill-seekers cloud upload output/ --provider s3 --bucket my-bucket
Auto-update your AI knowledge on every doc change.
---
Tweet 9: Get Started
pip install skill-seekers
# Try an example
skill-seekers scrape --config configs/react.json
# Or create your own
skill-seekers config --wizard
---
Tweet 10: Links
🌐 Website: https://skillseekersweb.com
💻 GitHub: https://github.com/yusufkaraaslan/Skill_Seekers
📖 Docs: https://skillseekersweb.com/docs
Star ⭐ if you hate writing scrapers.
#AI #RAG #LangChain #OpenSource
Afternoon (1-3 PM EST)
LinkedIn Post (Professional angle)
🚀 Launching Skill Seekers v3.0.0
After months of development, we're launching the universal
documentation preprocessor for AI systems.
What started as a Claude skill generator has evolved into
a platform that serves the entire AI ecosystem:
✅ 16 output formats (LangChain, LlamaIndex, Pinecone, Cursor, etc.)
✅ 26 MCP tools for AI agents
✅ Cloud storage (S3, GCS, Azure)
✅ CI/CD ready (GitHub Action + Docker)
✅ 1,852 tests, production-ready
The problem we solve: Every AI team spends weeks building
documentation scrapers. We eliminate that entirely.
One command. Any format. Production-ready.
Try it: pip install skill-seekers
#AI #MachineLearning #DeveloperTools #OpenSource #RAG
Evening
- Respond to all comments/questions
- Retweet with additional insights
- Share in relevant Discord/Slack communities
DAY 3: Reddit & Communities (Wednesday)
Theme: "Community Engagement"
Morning (9-11 AM EST)
Post 1: r/LangChain
Title: "Skill Seekers v3.0.0 - Universal preprocessor now supports LangChain Documents"
Hey r/LangChain!
We just launched v3.0.0 of Skill Seekers, and it now outputs
LangChain Document objects directly.
What it does:
- Scrapes documentation websites
- Preserves code blocks (doesn't split them)
- Adds rich metadata (source, category, url)
- Outputs LangChain Documents ready for vector stores
Example:
```python
# CLI
skill-seekers scrape --format langchain --config react.json
# Python
from skill_seekers.cli.adaptors import get_adaptor
adaptor = get_adaptor('langchain')
documents = adaptor.load_documents("output/react/")
# Now use with any LangChain vector store
Key features:
- 16 output formats total
- 1,852 tests passing
- 26 MCP tools
- Works with Chroma, Pinecone, Weaviate, Qdrant, FAISS
GitHub: [link] Website: [link]
Would love your feedback!
**Post 2: r/cursor**
Title: "Give Cursor complete framework knowledge with Skill Seekers v3.0.0"
Cursor users - tired of generic suggestions?
We built a tool that converts any framework documentation into .cursorrules files.
Example - React:
skill-seekers scrape --target claude --config react.json
cp output/react-claude/.cursorrules ./
Result: Cursor now knows React hooks, patterns, best practices.
Before: Generic "useState" suggestions After: "Consider using useReducer for complex state logic" with examples
Also works for:
- Vue, Angular, Svelte
- Django, FastAPI, Rails
- Any framework with docs
v3.0.0 adds support for:
- Windsurf (.windsurfrules)
- Cline (.clinerules)
- Continue.dev
Try it: pip install skill-seekers
GitHub: [link]
**Post 3: r/LLMDevs**
Title: "Skill Seekers v3.0.0 - The universal documentation preprocessor (16 formats, 1,852 tests)"
TL;DR: One tool converts docs into any AI format.
Formats supported:
- RAG: LangChain, LlamaIndex, Haystack, Pinecone-ready
- Vector DBs: Chroma, Weaviate, Qdrant, FAISS
- AI Coding: Cursor, Windsurf, Cline, Continue.dev
- AI Platforms: Claude, Gemini, OpenAI
- Generic: Markdown
MCP Tools: 26 tools for AI agents Cloud: S3, GCS, Azure CI/CD: GitHub Action, Docker
Stats:
- 58,512 LOC
- 1,852 tests
- 100 test files
- 12 example projects
The pitch: Stop rebuilding doc scrapers. Use this.
pip install skill-seekers
GitHub: [link] Website: [link]
AMA!
#### Afternoon (1-3 PM EST)
**Hacker News - Show HN**
Title: "Show HN: Skill Seekers v3.0.0 – Universal doc preprocessor for AI systems"
We built a tool that transforms documentation into structured knowledge for any AI system.
Problem: Every AI project needs documentation, but everyone rebuilds the same scrapers.
Solution: One command → 16 output formats
Supported:
- RAG: LangChain, LlamaIndex, Haystack
- Vector DBs: Chroma, Weaviate, Qdrant, FAISS
- AI Coding: Cursor, Windsurf, Cline, Continue.dev
- AI Platforms: Claude, Gemini, OpenAI
Tech stack:
- Python 3.10+
- 1,852 tests
- MCP (Model Context Protocol)
- GitHub Action + Docker
Examples:
# LangChain
skill-seekers scrape --format langchain --config react.json
# Cursor
skill-seekers scrape --target claude --config react.json
# Direct to cloud
skill-seekers cloud upload output/ --provider s3 --bucket my-bucket
Website: https://skillseekersweb.com GitHub: https://github.com/yusufkaraaslan/Skill_Seekers
Would love feedback from the HN community!
#### Evening
- [ ] Respond to ALL comments
- [ ] Upvote helpful responses
- [ ] Cross-reference between posts
---
### DAY 4: Partnership Outreach (Thursday)
**Theme:** "Partnership Push"
#### Morning (9-11 AM EST)
**Send 6 emails simultaneously:**
1. **LangChain** (contact@langchain.dev)
2. **LlamaIndex** (hello@llamaindex.ai)
3. **Pinecone** (community@pinecone.io)
4. **Cursor** (support@cursor.sh)
5. **Windsurf** (hello@codeium.com)
6. **Cline** (via GitHub/Twitter @saoudrizwan)
**Email Template:**
Subject: Skill Seekers v3.0.0 - Official [Platform] Integration + Partnership
Hi [Name/Team],
We just launched Skill Seekers v3.0.0 with official [Platform] integration, and I'd love to explore a partnership.
What we built:
- [Platform] integration: [specific details]
- Working example: [link to example in our repo]
- Integration guide: [link]
We have:
- 12 complete example projects
- 18 integration guides
- 1,852 tests, production-ready
- Active community
What we'd love:
- Mention in your docs/examples
- Feedback on the integration
- Potential collaboration
Demo: [link to working example]
Best, [Your Name] Skill Seekers https://skillseekersweb.com/
#### Afternoon (1-3 PM EST)
- [ ] **Product Hunt Submission**
- Title: "Skill Seekers v3.0.0"
- Tagline: "Universal documentation preprocessor for AI systems"
- Category: Developer Tools
- Images: Screenshots of different formats
- [ ] **Indie Hackers Post**
- Share launch story
- Technical challenges
- Lessons learned
#### Evening
- [ ] Check email responses
- [ ] Follow up on social engagement
---
### DAY 5: Content & Examples (Friday)
**Theme:** "Deep Dive Content"
#### Morning (9-11 AM EST)
**Publish RAG Tutorial Blog Post**
Title: "From Documentation to RAG Pipeline in 5 Minutes"
Step-by-step tutorial:
- Scrape React docs
- Convert to LangChain Documents
- Store in Chroma
- Query with natural language
Complete code included.
**Publish AI Coding Guide**
Title: "Give Cursor Complete Framework Knowledge"
Before/after comparison:
- Without: Generic suggestions
- With: Framework-specific intelligence
Covers: Cursor, Windsurf, Cline, Continue.dev
#### Afternoon (1-3 PM EST)
**YouTube/Video Platforms** (if applicable)
- Create 2-minute demo video
- Post on YouTube, TikTok, Instagram Reels
**Newsletter/Email List** (if you have one)
- Send launch announcement to subscribers
#### Evening
- [ ] Compile Week 1 metrics
- [ ] Plan follow-up content
- [ ] Respond to all remaining comments
---
## 📊 WEEKEND: Monitor & Engage
### Saturday-Sunday
- [ ] Monitor all platforms for comments
- [ ] Respond within 2 hours to everything
- [ ] Share best comments/testimonials
- [ ] Prepare Week 2 follow-up content
---
## 🎯 CONTENT CALENDAR AT A GLANCE
| Day | Platform | Content | Time |
|-----|----------|---------|------|
| **Mon** | PyPI, GitHub | Release | Morning |
| | Website | Blog post | Afternoon |
| **Tue** | Twitter | 10-tweet thread | Morning |
| | LinkedIn | Professional post | Afternoon |
| **Wed** | Reddit | 3 posts (r/LangChain, r/cursor, r/LLMDevs) | Morning |
| | HN | Show HN | Afternoon |
| **Thu** | Email | 6 partnership emails | Morning |
| | Product Hunt | Submission | Afternoon |
| **Fri** | Website | 2 blog posts (tutorial + guide) | Morning |
| | Video | Demo video | Afternoon |
| **Weekend** | All | Monitor & engage | Ongoing |
---
## 📈 SUCCESS METRICS (5 Days)
| Metric | Conservative | Target | Stretch |
|--------|-------------|--------|---------|
| **GitHub Stars** | +50 | +75 | +100 |
| **PyPI Downloads** | +300 | +500 | +800 |
| **Blog Views** | 1,500 | 2,500 | 4,000 |
| **Social Engagement** | 100 | 250 | 500 |
| **Email Responses** | 2 | 4 | 6 |
| **HN Upvotes** | 50 | 100 | 200 |
---
## 🚀 WHY THIS WORKS BETTER
### 4-Week Approach Problems:
- ❌ Momentum dies between weeks
- ❌ People forget after first week
- ❌ Harder to coordinate multiple channels
- ❌ Competitors might launch similar
### 1-Week Blitz Advantages:
- ✅ Creates "surge" effect - everywhere at once
- ✅ Easier to coordinate and track
- ✅ Builds on momentum day by day
- ✅ Faster feedback loop
- ✅ Gets it DONE (vs. dragging out)
---
## ✅ PRE-LAUNCH CHECKLIST (Do Today)
- [ ] PyPI account ready
- [ ] Dev.to account created
- [ ] Twitter ready
- [ ] LinkedIn ready
- [ ] Reddit account (7+ days old)
- [ ] Hacker News account
- [ ] Product Hunt account
- [ ] All content reviewed
- [ ] Website live and tested
- [ ] Examples working
---
## 🎬 START NOW
**Your 3 actions for TODAY:**
1. **Publish to PyPI** (15 min)
2. **Create GitHub Release** (10 min)
3. **Schedule/publish first blog post** (30 min)
**Tomorrow:** Twitter thread + LinkedIn
**Wednesday:** Reddit + Hacker News
**Thursday:** Partnership emails
**Friday:** Tutorial content
---
**All-in-one week. Maximum impact. Let's GO! 🚀**