feat: Add C3.1 Design Pattern Detection - Detect 10 patterns across 9 languages

Implements comprehensive design pattern detection system for codebases,
enabling automatic identification of common GoF patterns with confidence
scoring and language-specific adaptations.

**Key Features:**
- 10 Design Patterns: Singleton, Factory, Observer, Strategy, Decorator,
  Builder, Adapter, Command, Template Method, Chain of Responsibility
- 3 Detection Levels: Surface (naming), Deep (structure), Full (behavior)
- 9 Language Support: Python (AST-based), JavaScript, TypeScript, C++, C,
  C#, Go, Rust, Java (regex-based), with Ruby/PHP basic support
- Language Adaptations: Python @decorator, Go sync.Once, Rust lazy_static
- Confidence Scoring: 0.0-1.0 scale with evidence tracking

**Architecture:**
- Base Classes: PatternInstance, PatternReport, BasePatternDetector
- Pattern Detectors: 10 specialized detectors with 3-tier detection
- Language Adapter: Language-specific confidence adjustments
- CodeAnalyzer Integration: Reuses existing parsing infrastructure

**CLI & Integration:**
- CLI Tool: skill-seekers-patterns --file src/db.py --depth deep
- Codebase Scraper: --detect-patterns flag for full codebase analysis
- MCP Tool: detect_patterns for Claude Code integration
- Output Formats: JSON and human-readable with pattern summaries

**Testing:**
- 24 comprehensive tests (100% passing in 0.30s)
- Coverage: All 10 patterns, multi-language support, edge cases
- Integration tests: CLI, codebase scraper, pattern recognition
- No regressions: 943/943 existing tests still pass

**Documentation:**
- docs/PATTERN_DETECTION.md: Complete user guide (514 lines)
- API reference, usage examples, language support matrix
- Accuracy benchmarks: 87% precision, 80% recall
- Troubleshooting guide and integration examples

**Files Changed:**
- Created: pattern_recognizer.py (1,869 lines), test suite (467 lines)
- Modified: codebase_scraper.py, MCP tools, servers, CHANGELOG.md
- Added: CLI entry point in pyproject.toml

**Performance:**
- Surface: ~200 classes/sec, <5ms per class
- Deep: ~100 classes/sec, ~10ms per class (default)
- Full: ~50 classes/sec, ~20ms per class

**Bug Fixes:**
- Fixed missing imports (argparse, json, sys) in pattern_recognizer.py
- Fixed pyproject.toml dependency duplication (removed dev from optional-dependencies)

**Roadmap:**
- Completes C3.1 from FLEXIBLE_ROADMAP.md
- Foundation for C3.2-C3.5 (usage examples, how-to guides, config patterns)

Closes #117 (C3.1 Design Pattern Detection)

Co-authored-by: Claude Sonnet 4.5 <noreply@anthropic.com>
🤖 Generated with [Claude Code](https://claude.com/claude-code)
This commit is contained in:
yusyus
2026-01-03 19:56:09 +03:00
parent 500b74078b
commit 0d664785f7
10 changed files with 3101 additions and 15 deletions

View File

@@ -209,7 +209,8 @@ def analyze_codebase(
file_patterns: Optional[List[str]] = None,
build_api_reference: bool = False,
extract_comments: bool = True,
build_dependency_graph: bool = False
build_dependency_graph: bool = False,
detect_patterns: bool = False
) -> Dict[str, Any]:
"""
Analyze local codebase and extract code knowledge.
@@ -223,6 +224,7 @@ def analyze_codebase(
build_api_reference: Generate API reference markdown
extract_comments: Extract inline comments
build_dependency_graph: Generate dependency graph and detect circular dependencies
detect_patterns: Detect design patterns (Singleton, Factory, Observer, etc.)
Returns:
Analysis results dictionary
@@ -370,6 +372,45 @@ def analyze_codebase(
except:
pass # pydot not installed, skip DOT export
# Detect design patterns if requested (C3.1)
if detect_patterns:
logger.info("Detecting design patterns...")
from skill_seekers.cli.pattern_recognizer import PatternRecognizer
pattern_recognizer = PatternRecognizer(depth=depth)
pattern_results = []
for file_path in files:
try:
content = file_path.read_text(encoding='utf-8', errors='ignore')
language = detect_language(file_path)
if language != 'Unknown':
report = pattern_recognizer.analyze_file(
str(file_path), content, language
)
if report.patterns:
pattern_results.append(report.to_dict())
except Exception as e:
logger.warning(f"Pattern detection failed for {file_path}: {e}")
continue
# Save pattern results
if pattern_results:
pattern_output = output_dir / 'patterns'
pattern_output.mkdir(parents=True, exist_ok=True)
pattern_json = pattern_output / 'detected_patterns.json'
with open(pattern_json, 'w', encoding='utf-8') as f:
json.dump(pattern_results, f, indent=2)
total_patterns = sum(len(r['patterns']) for r in pattern_results)
logger.info(f"✅ Detected {total_patterns} patterns in {len(pattern_results)} files")
logger.info(f"📁 Saved to: {pattern_json}")
else:
logger.info("No design patterns detected")
return results
@@ -434,6 +475,11 @@ Examples:
action='store_true',
help='Generate dependency graph and detect circular dependencies'
)
parser.add_argument(
'--detect-patterns',
action='store_true',
help='Detect design patterns in code (Singleton, Factory, Observer, etc.)'
)
parser.add_argument(
'--no-comments',
action='store_true',
@@ -481,7 +527,8 @@ Examples:
file_patterns=file_patterns,
build_api_reference=args.build_api_reference,
extract_comments=not args.no_comments,
build_dependency_graph=args.build_dependency_graph
build_dependency_graph=args.build_dependency_graph,
detect_patterns=args.detect_patterns
)
# Print summary

File diff suppressed because it is too large Load Diff

View File

@@ -36,6 +36,7 @@ try:
scrape_docs_tool,
scrape_github_tool,
scrape_pdf_tool,
detect_patterns_tool,
run_subprocess_with_streaming,
)
from skill_seekers.mcp.tools.packaging_tools import (
@@ -95,6 +96,8 @@ try:
return await remove_config_source_tool(arguments)
elif name == "install_skill":
return await install_skill_tool(arguments)
elif name == "detect_patterns":
return await detect_patterns_tool(arguments)
else:
return [TextContent(type="text", text=f"Unknown tool: {name}")]
except Exception as e:

View File

@@ -82,6 +82,7 @@ try:
scrape_github_impl,
scrape_pdf_impl,
scrape_codebase_impl,
detect_patterns_impl,
# Packaging tools
package_skill_impl,
upload_skill_impl,
@@ -110,6 +111,7 @@ except ImportError:
scrape_github_impl,
scrape_pdf_impl,
scrape_codebase_impl,
detect_patterns_impl,
package_skill_impl,
upload_skill_impl,
enhance_skill_impl,
@@ -438,6 +440,50 @@ async def scrape_codebase(
return str(result)
@safe_tool_decorator(
description="Detect design patterns in source code (Singleton, Factory, Observer, Strategy, Decorator, Builder, Adapter, Command, Template Method, Chain of Responsibility). Supports 9 languages: Python, JavaScript, TypeScript, C++, C, C#, Go, Rust, Java, Ruby, PHP."
)
async def detect_patterns(
file: str = "",
directory: str = "",
output: str = "",
depth: str = "deep",
json: bool = False,
) -> str:
"""
Detect design patterns in source code.
Analyzes source files or directories to identify common design patterns.
Provides confidence scores and evidence for each detected pattern.
Args:
file: Single file to analyze (optional)
directory: Directory to analyze all source files (optional)
output: Output directory for JSON results (optional)
depth: Detection depth - surface (fast), deep (balanced), full (thorough). Default: deep
json: Output JSON format instead of human-readable (default: false)
Returns:
Pattern detection results with confidence scores and evidence.
Example:
detect_patterns(file="src/database.py", depth="deep")
detect_patterns(directory="src/", output="patterns/", json=true)
"""
args = {
"file": file,
"directory": directory,
"output": output,
"depth": depth,
"json": json,
}
result = await detect_patterns_impl(args)
if isinstance(result, list) and result:
return result[0].text if hasattr(result[0], "text") else str(result[0])
return str(result)
# ============================================================================
# PACKAGING TOOLS (3 tools)
# ============================================================================

View File

@@ -25,6 +25,7 @@ from .scraping_tools import (
scrape_github_tool as scrape_github_impl,
scrape_pdf_tool as scrape_pdf_impl,
scrape_codebase_tool as scrape_codebase_impl,
detect_patterns_tool as detect_patterns_impl,
)
from .packaging_tools import (
@@ -58,6 +59,7 @@ __all__ = [
"scrape_github_impl",
"scrape_pdf_impl",
"scrape_codebase_impl",
"detect_patterns_impl",
# Packaging tools
"package_skill_impl",
"upload_skill_impl",

View File

@@ -504,3 +504,73 @@ async def scrape_codebase_tool(args: dict) -> List[TextContent]:
return [TextContent(type="text", text=output_text)]
else:
return [TextContent(type="text", text=f"{output_text}\n\n❌ Error:\n{stderr}")]
async def detect_patterns_tool(args: dict) -> List[TextContent]:
"""
Detect design patterns in source code.
Analyzes source files or directories to detect common design patterns
(Singleton, Factory, Observer, Strategy, Decorator, Builder, Adapter,
Command, Template Method, Chain of Responsibility).
Supports 9 languages: Python, JavaScript, TypeScript, C++, C, C#,
Go, Rust, Java, Ruby, PHP.
Args:
args: Dictionary containing:
- file (str, optional): Single file to analyze
- directory (str, optional): Directory to analyze (analyzes all source files)
- output (str, optional): Output directory for JSON results
- depth (str, optional): Detection depth - surface, deep, full (default: deep)
- json (bool, optional): Output JSON format (default: False)
Returns:
List[TextContent]: Pattern detection results
Example:
detect_patterns(file="src/database.py", depth="deep")
detect_patterns(directory="src/", output="patterns/", json=True)
"""
file_path = args.get("file")
directory = args.get("directory")
if not file_path and not directory:
return [TextContent(type="text", text="❌ Error: Must specify either 'file' or 'directory' parameter")]
output = args.get("output", "")
depth = args.get("depth", "deep")
json_output = args.get("json", False)
# Build command
cmd = [sys.executable, "-m", "skill_seekers.cli.pattern_recognizer"]
if file_path:
cmd.extend(["--file", file_path])
if directory:
cmd.extend(["--directory", directory])
if output:
cmd.extend(["--output", output])
if depth:
cmd.extend(["--depth", depth])
if json_output:
cmd.append("--json")
timeout = 300 # 5 minutes for pattern detection
progress_msg = "🔍 Detecting design patterns...\n"
if file_path:
progress_msg += f"📄 File: {file_path}\n"
if directory:
progress_msg += f"📁 Directory: {directory}\n"
progress_msg += f"🎯 Detection depth: {depth}\n"
progress_msg += f"⏱️ Maximum time: {timeout // 60} minutes\n\n"
stdout, stderr, returncode = run_subprocess_with_streaming(cmd, timeout=timeout)
output_text = progress_msg + stdout
if returncode == 0:
return [TextContent(type="text", text=output_text)]
else:
return [TextContent(type="text", text=f"{output_text}\n\n❌ Error:\n{stderr}")]