Test Framework: - Created tests/ directory structure - Added __init__.py for test package - Implemented 71 comprehensive tests across 3 test suites Test Suites: 1. test_config_validation.py (25 tests) - Valid/invalid config structure - Required fields validation - Name format validation - URL format validation - Selectors validation - URL patterns validation - Categories validation - Rate limit validation (0-10 range) - Max pages validation (1-10000 range) - Start URLs validation 2. test_scraper_features.py (28 tests) - URL validation (include/exclude patterns) - Language detection (Python, JavaScript, GDScript, C++, etc.) - Pattern extraction from documentation - Smart categorization (by URL, title, content) - Text cleaning utilities 3. test_integration.py (18 tests) - Dry-run mode functionality - Config loading and validation - Real config files validation (godot, react, vue, django, fastapi, steam) - URL processing and normalization - Content extraction Test Runner (run_tests.py): - Custom colored test runner with ANSI colors - Detailed test summary with breakdown by category - Success rate calculation - Command-line options: --suite: Run specific test suite --verbose: Show each test name --quiet: Minimal output --failfast: Stop on first failure --list: List all available tests - Execution time: ~1 second for full suite Documentation: - Added comprehensive TESTING.md guide - Test writing templates - Best practices - Coverage information - Troubleshooting guide .gitignore: - Added Python cache files - Added output directory - Added IDE and OS files Test Results: ✅ 71/71 tests passing (100% pass rate) ✅ All existing configs validated ✅ Fast execution (<1 second) ✅ Ready for CI/CD integration 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude <noreply@anthropic.com>
229 lines
7.2 KiB
Python
Executable File
229 lines
7.2 KiB
Python
Executable File
#!/usr/bin/env python3
|
|
"""
|
|
Test Runner for Skill Seeker
|
|
Runs all test suites and generates a comprehensive test report
|
|
"""
|
|
|
|
import sys
|
|
import unittest
|
|
import os
|
|
from io import StringIO
|
|
from pathlib import Path
|
|
|
|
|
|
class ColoredTextTestResult(unittest.TextTestResult):
|
|
"""Custom test result class with colored output"""
|
|
|
|
# ANSI color codes
|
|
GREEN = '\033[92m'
|
|
RED = '\033[91m'
|
|
YELLOW = '\033[93m'
|
|
BLUE = '\033[94m'
|
|
RESET = '\033[0m'
|
|
BOLD = '\033[1m'
|
|
|
|
def __init__(self, *args, **kwargs):
|
|
super().__init__(*args, **kwargs)
|
|
self.test_results = []
|
|
|
|
def addSuccess(self, test):
|
|
super().addSuccess(test)
|
|
self.test_results.append(('PASS', test))
|
|
if self.showAll:
|
|
self.stream.write(f"{self.GREEN}✓ PASS{self.RESET}\n")
|
|
elif self.dots:
|
|
self.stream.write(f"{self.GREEN}.{self.RESET}")
|
|
self.stream.flush()
|
|
|
|
def addError(self, test, err):
|
|
super().addError(test, err)
|
|
self.test_results.append(('ERROR', test))
|
|
if self.showAll:
|
|
self.stream.write(f"{self.RED}✗ ERROR{self.RESET}\n")
|
|
elif self.dots:
|
|
self.stream.write(f"{self.RED}E{self.RESET}")
|
|
self.stream.flush()
|
|
|
|
def addFailure(self, test, err):
|
|
super().addFailure(test, err)
|
|
self.test_results.append(('FAIL', test))
|
|
if self.showAll:
|
|
self.stream.write(f"{self.RED}✗ FAIL{self.RESET}\n")
|
|
elif self.dots:
|
|
self.stream.write(f"{self.RED}F{self.RESET}")
|
|
self.stream.flush()
|
|
|
|
def addSkip(self, test, reason):
|
|
super().addSkip(test, reason)
|
|
self.test_results.append(('SKIP', test))
|
|
if self.showAll:
|
|
self.stream.write(f"{self.YELLOW}⊘ SKIP{self.RESET}\n")
|
|
elif self.dots:
|
|
self.stream.write(f"{self.YELLOW}s{self.RESET}")
|
|
self.stream.flush()
|
|
|
|
|
|
class ColoredTextTestRunner(unittest.TextTestRunner):
|
|
"""Custom test runner with colored output"""
|
|
resultclass = ColoredTextTestResult
|
|
|
|
|
|
def discover_tests(test_dir='tests'):
|
|
"""Discover all test files in the tests directory"""
|
|
loader = unittest.TestLoader()
|
|
start_dir = test_dir
|
|
pattern = 'test_*.py'
|
|
|
|
suite = loader.discover(start_dir, pattern=pattern)
|
|
return suite
|
|
|
|
|
|
def run_specific_suite(suite_name):
|
|
"""Run a specific test suite"""
|
|
loader = unittest.TestLoader()
|
|
|
|
suite_map = {
|
|
'config': 'tests.test_config_validation',
|
|
'features': 'tests.test_scraper_features',
|
|
'integration': 'tests.test_integration'
|
|
}
|
|
|
|
if suite_name not in suite_map:
|
|
print(f"Unknown test suite: {suite_name}")
|
|
print(f"Available suites: {', '.join(suite_map.keys())}")
|
|
return None
|
|
|
|
module_name = suite_map[suite_name]
|
|
try:
|
|
suite = loader.loadTestsFromName(module_name)
|
|
return suite
|
|
except Exception as e:
|
|
print(f"Error loading test suite '{suite_name}': {e}")
|
|
return None
|
|
|
|
|
|
def print_summary(result):
|
|
"""Print a detailed test summary"""
|
|
total = result.testsRun
|
|
passed = total - len(result.failures) - len(result.errors) - len(result.skipped)
|
|
failed = len(result.failures)
|
|
errors = len(result.errors)
|
|
skipped = len(result.skipped)
|
|
|
|
print("\n" + "="*70)
|
|
print("TEST SUMMARY")
|
|
print("="*70)
|
|
|
|
# Overall stats
|
|
print(f"\n{ColoredTextTestResult.BOLD}Total Tests:{ColoredTextTestResult.RESET} {total}")
|
|
print(f"{ColoredTextTestResult.GREEN}✓ Passed:{ColoredTextTestResult.RESET} {passed}")
|
|
if failed > 0:
|
|
print(f"{ColoredTextTestResult.RED}✗ Failed:{ColoredTextTestResult.RESET} {failed}")
|
|
if errors > 0:
|
|
print(f"{ColoredTextTestResult.RED}✗ Errors:{ColoredTextTestResult.RESET} {errors}")
|
|
if skipped > 0:
|
|
print(f"{ColoredTextTestResult.YELLOW}⊘ Skipped:{ColoredTextTestResult.RESET} {skipped}")
|
|
|
|
# Success rate
|
|
if total > 0:
|
|
success_rate = (passed / total) * 100
|
|
color = ColoredTextTestResult.GREEN if success_rate == 100 else \
|
|
ColoredTextTestResult.YELLOW if success_rate >= 80 else \
|
|
ColoredTextTestResult.RED
|
|
print(f"\n{color}Success Rate: {success_rate:.1f}%{ColoredTextTestResult.RESET}")
|
|
|
|
# Category breakdown
|
|
if hasattr(result, 'test_results'):
|
|
print(f"\n{ColoredTextTestResult.BOLD}Test Breakdown by Category:{ColoredTextTestResult.RESET}")
|
|
|
|
categories = {}
|
|
for status, test in result.test_results:
|
|
test_name = str(test)
|
|
# Extract test class name
|
|
if '.' in test_name:
|
|
class_name = test_name.split('.')[0].split()[-1]
|
|
if class_name not in categories:
|
|
categories[class_name] = {'PASS': 0, 'FAIL': 0, 'ERROR': 0, 'SKIP': 0}
|
|
categories[class_name][status] += 1
|
|
|
|
for category, stats in sorted(categories.items()):
|
|
total_cat = sum(stats.values())
|
|
passed_cat = stats['PASS']
|
|
print(f" {category}: {passed_cat}/{total_cat} passed")
|
|
|
|
print("\n" + "="*70)
|
|
|
|
# Return status
|
|
return failed == 0 and errors == 0
|
|
|
|
|
|
def main():
|
|
"""Main test runner"""
|
|
import argparse
|
|
|
|
parser = argparse.ArgumentParser(
|
|
description='Run tests for Skill Seeker',
|
|
formatter_class=argparse.RawDescriptionHelpFormatter
|
|
)
|
|
|
|
parser.add_argument('--suite', '-s', type=str,
|
|
help='Run specific test suite (config, features, integration)')
|
|
parser.add_argument('--verbose', '-v', action='store_true',
|
|
help='Verbose output (show each test)')
|
|
parser.add_argument('--quiet', '-q', action='store_true',
|
|
help='Quiet output (minimal output)')
|
|
parser.add_argument('--failfast', '-f', action='store_true',
|
|
help='Stop on first failure')
|
|
parser.add_argument('--list', '-l', action='store_true',
|
|
help='List all available tests')
|
|
|
|
args = parser.parse_args()
|
|
|
|
# Set verbosity
|
|
verbosity = 1
|
|
if args.verbose:
|
|
verbosity = 2
|
|
elif args.quiet:
|
|
verbosity = 0
|
|
|
|
print(f"\n{ColoredTextTestResult.BOLD}{'='*70}{ColoredTextTestResult.RESET}")
|
|
print(f"{ColoredTextTestResult.BOLD}SKILL SEEKER TEST SUITE{ColoredTextTestResult.RESET}")
|
|
print(f"{ColoredTextTestResult.BOLD}{'='*70}{ColoredTextTestResult.RESET}\n")
|
|
|
|
# Discover or load specific suite
|
|
if args.suite:
|
|
print(f"Running test suite: {ColoredTextTestResult.BLUE}{args.suite}{ColoredTextTestResult.RESET}\n")
|
|
suite = run_specific_suite(args.suite)
|
|
if suite is None:
|
|
return 1
|
|
else:
|
|
print(f"Running {ColoredTextTestResult.BLUE}all tests{ColoredTextTestResult.RESET}\n")
|
|
suite = discover_tests()
|
|
|
|
# List tests
|
|
if args.list:
|
|
print("\nAvailable tests:\n")
|
|
for test_group in suite:
|
|
for test in test_group:
|
|
print(f" - {test}")
|
|
print()
|
|
return 0
|
|
|
|
# Run tests
|
|
runner = ColoredTextTestRunner(
|
|
verbosity=verbosity,
|
|
failfast=args.failfast
|
|
)
|
|
|
|
result = runner.run(suite)
|
|
|
|
# Print summary
|
|
success = print_summary(result)
|
|
|
|
# Return appropriate exit code
|
|
return 0 if success else 1
|
|
|
|
|
|
if __name__ == '__main__':
|
|
sys.exit(main())
|