Files
claude-skills-reference/engineering-team/senior-qa/scripts/coverage_analyzer.py
Alireza Rezvani 6cd35fedd8 fix(skill): rewrite senior-qa with unique, actionable content (#51) (#95)
Complete rewrite of the senior-qa skill addressing all feedback from Issue #51:

SKILL.md (444 lines):
- Added proper YAML frontmatter with trigger phrases
- Added Table of Contents
- Focused on React/Next.js testing (Jest, RTL, Playwright)
- 3 actionable workflows with numbered steps
- Removed marketing language

References (3 files, 2,625+ lines total):
- testing_strategies.md: Test pyramid, coverage targets, CI/CD patterns
- test_automation_patterns.md: Page Object Model, fixtures, mocking, async testing
- qa_best_practices.md: Naming conventions, isolation, debugging strategies

Scripts (3 files, 2,261+ lines total):
- test_suite_generator.py: Scans React components, generates Jest+RTL tests
- coverage_analyzer.py: Parses Istanbul/LCOV, identifies critical gaps
- e2e_test_scaffolder.py: Scans Next.js routes, generates Playwright tests

Documentation:
- Updated engineering-team/README.md senior-qa section
- Added README.md in senior-qa subfolder

Resolves #51

Co-authored-by: Claude Opus 4.5 <noreply@anthropic.com>
2026-01-27 08:25:56 +01:00

837 lines
29 KiB
Python
Executable File

#!/usr/bin/env python3
"""
Coverage Analyzer
Parses Jest/Istanbul coverage reports and identifies gaps, uncovered branches,
and provides actionable recommendations for improving test coverage.
Usage:
python coverage_analyzer.py coverage/coverage-final.json --threshold 80
python coverage_analyzer.py coverage/ --format html --output report.html
python coverage_analyzer.py coverage/ --critical-paths
"""
import os
import sys
import json
import argparse
import re
from pathlib import Path
from typing import Dict, List, Optional, Tuple, Any
from dataclasses import dataclass, field, asdict
from datetime import datetime
from collections import defaultdict
@dataclass
class FileCoverage:
"""Coverage data for a single file"""
path: str
statements: Tuple[int, int] # (covered, total)
branches: Tuple[int, int]
functions: Tuple[int, int]
lines: Tuple[int, int]
uncovered_lines: List[int] = field(default_factory=list)
uncovered_branches: List[str] = field(default_factory=list)
@property
def statement_pct(self) -> float:
return (self.statements[0] / self.statements[1] * 100) if self.statements[1] > 0 else 100
@property
def branch_pct(self) -> float:
return (self.branches[0] / self.branches[1] * 100) if self.branches[1] > 0 else 100
@property
def function_pct(self) -> float:
return (self.functions[0] / self.functions[1] * 100) if self.functions[1] > 0 else 100
@property
def line_pct(self) -> float:
return (self.lines[0] / self.lines[1] * 100) if self.lines[1] > 0 else 100
@dataclass
class CoverageGap:
"""An identified coverage gap"""
file: str
gap_type: str # 'statements', 'branches', 'functions', 'lines'
lines: List[int]
severity: str # 'critical', 'high', 'medium', 'low'
description: str
recommendation: str
@dataclass
class CoverageSummary:
"""Overall coverage summary"""
statements: Tuple[int, int]
branches: Tuple[int, int]
functions: Tuple[int, int]
lines: Tuple[int, int]
files_analyzed: int
files_below_threshold: int = 0
class CoverageParser:
"""Parses various coverage report formats"""
def __init__(self, verbose: bool = False):
self.verbose = verbose
def parse(self, path: Path) -> Tuple[Dict[str, FileCoverage], CoverageSummary]:
"""Parse coverage data from file or directory"""
if path.is_file():
if path.suffix == '.json':
return self._parse_istanbul_json(path)
elif path.suffix == '.info' or 'lcov' in path.name:
return self._parse_lcov(path)
elif path.is_dir():
# Look for common coverage files
for filename in ['coverage-final.json', 'coverage-summary.json', 'lcov.info']:
candidate = path / filename
if candidate.exists():
return self.parse(candidate)
# Check for coverage-final.json in coverage directory
coverage_json = path / 'coverage-final.json'
if coverage_json.exists():
return self._parse_istanbul_json(coverage_json)
raise ValueError(f"Could not find or parse coverage data at: {path}")
def _parse_istanbul_json(self, path: Path) -> Tuple[Dict[str, FileCoverage], CoverageSummary]:
"""Parse Istanbul/Jest JSON coverage format"""
with open(path, 'r') as f:
data = json.load(f)
files = {}
total_statements = [0, 0]
total_branches = [0, 0]
total_functions = [0, 0]
total_lines = [0, 0]
for file_path, file_data in data.items():
# Skip node_modules
if 'node_modules' in file_path:
continue
# Parse statement coverage
s_map = file_data.get('statementMap', {})
s_hits = file_data.get('s', {})
covered_statements = sum(1 for h in s_hits.values() if h > 0)
total_statements[0] += covered_statements
total_statements[1] += len(s_map)
# Parse branch coverage
b_map = file_data.get('branchMap', {})
b_hits = file_data.get('b', {})
covered_branches = sum(
sum(1 for h in hits if h > 0)
for hits in b_hits.values()
)
total_branch_count = sum(len(b['locations']) for b in b_map.values())
total_branches[0] += covered_branches
total_branches[1] += total_branch_count
# Parse function coverage
fn_map = file_data.get('fnMap', {})
fn_hits = file_data.get('f', {})
covered_functions = sum(1 for h in fn_hits.values() if h > 0)
total_functions[0] += covered_functions
total_functions[1] += len(fn_map)
# Determine uncovered lines
uncovered_lines = []
for stmt_id, hits in s_hits.items():
if hits == 0 and stmt_id in s_map:
stmt = s_map[stmt_id]
start_line = stmt.get('start', {}).get('line', 0)
if start_line not in uncovered_lines:
uncovered_lines.append(start_line)
# Count lines
line_coverage = self._calculate_line_coverage(s_map, s_hits)
total_lines[0] += line_coverage[0]
total_lines[1] += line_coverage[1]
# Identify uncovered branches
uncovered_branches = []
for branch_id, hits in b_hits.items():
for idx, hit in enumerate(hits):
if hit == 0:
uncovered_branches.append(f"{branch_id}:{idx}")
files[file_path] = FileCoverage(
path=file_path,
statements=(covered_statements, len(s_map)),
branches=(covered_branches, total_branch_count),
functions=(covered_functions, len(fn_map)),
lines=line_coverage,
uncovered_lines=sorted(uncovered_lines)[:50], # Limit
uncovered_branches=uncovered_branches[:20]
)
summary = CoverageSummary(
statements=tuple(total_statements),
branches=tuple(total_branches),
functions=tuple(total_functions),
lines=tuple(total_lines),
files_analyzed=len(files)
)
return files, summary
def _calculate_line_coverage(self, s_map: Dict, s_hits: Dict) -> Tuple[int, int]:
"""Calculate line coverage from statement data"""
lines = set()
covered_lines = set()
for stmt_id, stmt in s_map.items():
start_line = stmt.get('start', {}).get('line', 0)
end_line = stmt.get('end', {}).get('line', start_line)
for line in range(start_line, end_line + 1):
lines.add(line)
if s_hits.get(stmt_id, 0) > 0:
covered_lines.add(line)
return (len(covered_lines), len(lines))
def _parse_lcov(self, path: Path) -> Tuple[Dict[str, FileCoverage], CoverageSummary]:
"""Parse LCOV format coverage data"""
with open(path, 'r') as f:
content = f.read()
files = {}
current_file = None
current_data = {}
total = {
'statements': [0, 0],
'branches': [0, 0],
'functions': [0, 0],
'lines': [0, 0]
}
for line in content.split('\n'):
line = line.strip()
if line.startswith('SF:'):
current_file = line[3:]
current_data = {
'lines_hit': 0, 'lines_total': 0,
'functions_hit': 0, 'functions_total': 0,
'branches_hit': 0, 'branches_total': 0,
'uncovered_lines': []
}
elif line.startswith('DA:'):
parts = line[3:].split(',')
if len(parts) >= 2:
line_num = int(parts[0])
hits = int(parts[1])
current_data['lines_total'] += 1
if hits > 0:
current_data['lines_hit'] += 1
else:
current_data['uncovered_lines'].append(line_num)
elif line.startswith('FN:'):
current_data['functions_total'] += 1
elif line.startswith('FNDA:'):
parts = line[5:].split(',')
if len(parts) >= 1 and int(parts[0]) > 0:
current_data['functions_hit'] += 1
elif line.startswith('BRDA:'):
parts = line[5:].split(',')
current_data['branches_total'] += 1
if len(parts) >= 4 and parts[3] != '-' and int(parts[3]) > 0:
current_data['branches_hit'] += 1
elif line == 'end_of_record' and current_file:
# Skip node_modules
if 'node_modules' not in current_file:
files[current_file] = FileCoverage(
path=current_file,
statements=(current_data['lines_hit'], current_data['lines_total']),
branches=(current_data['branches_hit'], current_data['branches_total']),
functions=(current_data['functions_hit'], current_data['functions_total']),
lines=(current_data['lines_hit'], current_data['lines_total']),
uncovered_lines=current_data['uncovered_lines'][:50]
)
for key in total:
if key == 'statements' or key == 'lines':
total[key][0] += current_data['lines_hit']
total[key][1] += current_data['lines_total']
elif key == 'branches':
total[key][0] += current_data['branches_hit']
total[key][1] += current_data['branches_total']
elif key == 'functions':
total[key][0] += current_data['functions_hit']
total[key][1] += current_data['functions_total']
current_file = None
summary = CoverageSummary(
statements=tuple(total['statements']),
branches=tuple(total['branches']),
functions=tuple(total['functions']),
lines=tuple(total['lines']),
files_analyzed=len(files)
)
return files, summary
class CoverageAnalyzer:
"""Analyzes coverage data and generates recommendations"""
CRITICAL_PATTERNS = [
r'auth', r'payment', r'security', r'login', r'register',
r'checkout', r'order', r'transaction', r'billing'
]
SERVICE_PATTERNS = [
r'service', r'api', r'handler', r'controller', r'middleware'
]
def __init__(
self,
threshold: int = 80,
critical_paths: bool = False,
verbose: bool = False
):
self.threshold = threshold
self.critical_paths = critical_paths
self.verbose = verbose
def analyze(
self,
files: Dict[str, FileCoverage],
summary: CoverageSummary
) -> Tuple[List[CoverageGap], Dict[str, Any]]:
"""Analyze coverage and return gaps and recommendations"""
gaps = []
recommendations = {
'critical': [],
'high': [],
'medium': [],
'low': []
}
# Analyze each file
for file_path, coverage in files.items():
file_gaps = self._analyze_file(file_path, coverage)
gaps.extend(file_gaps)
# Sort gaps by severity
severity_order = {'critical': 0, 'high': 1, 'medium': 2, 'low': 3}
gaps.sort(key=lambda g: (severity_order[g.severity], -len(g.lines)))
# Generate recommendations
for gap in gaps:
recommendations[gap.severity].append({
'file': gap.file,
'type': gap.gap_type,
'lines': gap.lines[:10], # Limit
'description': gap.description,
'recommendation': gap.recommendation
})
# Add summary stats
stats = {
'overall_statement_pct': (summary.statements[0] / summary.statements[1] * 100) if summary.statements[1] > 0 else 100,
'overall_branch_pct': (summary.branches[0] / summary.branches[1] * 100) if summary.branches[1] > 0 else 100,
'overall_function_pct': (summary.functions[0] / summary.functions[1] * 100) if summary.functions[1] > 0 else 100,
'overall_line_pct': (summary.lines[0] / summary.lines[1] * 100) if summary.lines[1] > 0 else 100,
'files_analyzed': summary.files_analyzed,
'files_below_threshold': sum(
1 for f in files.values()
if f.line_pct < self.threshold
),
'total_gaps': len(gaps),
'critical_gaps': len(recommendations['critical']),
'threshold': self.threshold,
'meets_threshold': (summary.lines[0] / summary.lines[1] * 100) >= self.threshold if summary.lines[1] > 0 else True
}
return gaps, {
'recommendations': recommendations,
'stats': stats
}
def _analyze_file(self, file_path: str, coverage: FileCoverage) -> List[CoverageGap]:
"""Analyze a single file for coverage gaps"""
gaps = []
# Determine if file is critical
is_critical = any(
re.search(pattern, file_path.lower())
for pattern in self.CRITICAL_PATTERNS
)
is_service = any(
re.search(pattern, file_path.lower())
for pattern in self.SERVICE_PATTERNS
)
# Determine severity based on file type and coverage level
if is_critical:
base_severity = 'critical'
target_threshold = 95
elif is_service:
base_severity = 'high'
target_threshold = 85
else:
base_severity = 'medium'
target_threshold = self.threshold
# Check line coverage
if coverage.line_pct < target_threshold:
severity = base_severity if coverage.line_pct < 50 else self._lower_severity(base_severity)
gaps.append(CoverageGap(
file=file_path,
gap_type='lines',
lines=coverage.uncovered_lines[:20],
severity=severity,
description=f"Line coverage at {coverage.line_pct:.1f}% (target: {target_threshold}%)",
recommendation=self._get_line_recommendation(coverage)
))
# Check branch coverage
if coverage.branch_pct < target_threshold - 5: # Allow 5% less for branches
severity = base_severity if coverage.branch_pct < 40 else self._lower_severity(base_severity)
gaps.append(CoverageGap(
file=file_path,
gap_type='branches',
lines=[],
severity=severity,
description=f"Branch coverage at {coverage.branch_pct:.1f}%",
recommendation=f"Add tests for conditional logic. {len(coverage.uncovered_branches)} uncovered branches."
))
# Check function coverage
if coverage.function_pct < target_threshold:
severity = self._lower_severity(base_severity)
gaps.append(CoverageGap(
file=file_path,
gap_type='functions',
lines=[],
severity=severity,
description=f"Function coverage at {coverage.function_pct:.1f}%",
recommendation="Add tests for uncovered functions/methods."
))
return gaps
def _lower_severity(self, severity: str) -> str:
"""Lower severity by one level"""
mapping = {
'critical': 'high',
'high': 'medium',
'medium': 'low',
'low': 'low'
}
return mapping[severity]
def _get_line_recommendation(self, coverage: FileCoverage) -> str:
"""Generate recommendation for line coverage gaps"""
if coverage.line_pct < 30:
return "This file has very low coverage. Consider adding basic render/unit tests first."
elif coverage.line_pct < 60:
return "Add tests covering the main functionality and happy paths."
else:
return "Focus on edge cases and error handling paths."
class ReportGenerator:
"""Generates coverage reports in various formats"""
def __init__(self, verbose: bool = False):
self.verbose = verbose
def generate_text_report(
self,
files: Dict[str, FileCoverage],
summary: CoverageSummary,
analysis: Dict[str, Any],
threshold: int
) -> str:
"""Generate a text report"""
lines = []
# Header
lines.append("=" * 60)
lines.append("COVERAGE ANALYSIS REPORT")
lines.append(f"Generated: {datetime.now().strftime('%Y-%m-%d %H:%M:%S')}")
lines.append("=" * 60)
lines.append("")
# Overall summary
stats = analysis['stats']
lines.append("OVERALL COVERAGE:")
lines.append(f" Statements: {stats['overall_statement_pct']:.1f}%")
lines.append(f" Branches: {stats['overall_branch_pct']:.1f}%")
lines.append(f" Functions: {stats['overall_function_pct']:.1f}%")
lines.append(f" Lines: {stats['overall_line_pct']:.1f}%")
lines.append("")
# Threshold check
threshold_status = "PASS" if stats['meets_threshold'] else "FAIL"
lines.append(f"Threshold ({threshold}%): {threshold_status}")
lines.append(f"Files analyzed: {stats['files_analyzed']}")
lines.append(f"Files below threshold: {stats['files_below_threshold']}")
lines.append("")
# Critical gaps
recs = analysis['recommendations']
if recs['critical']:
lines.append("-" * 60)
lines.append("CRITICAL GAPS (requires immediate attention):")
for rec in recs['critical'][:5]:
lines.append(f" - {rec['file']}")
lines.append(f" {rec['description']}")
if rec['lines']:
lines.append(f" Uncovered lines: {', '.join(map(str, rec['lines'][:5]))}")
lines.append("")
# High priority gaps
if recs['high']:
lines.append("-" * 60)
lines.append("HIGH PRIORITY GAPS:")
for rec in recs['high'][:5]:
lines.append(f" - {rec['file']}")
lines.append(f" {rec['description']}")
lines.append("")
# Files below threshold
below_threshold = [
(path, cov) for path, cov in files.items()
if cov.line_pct < threshold
]
below_threshold.sort(key=lambda x: x[1].line_pct)
if below_threshold:
lines.append("-" * 60)
lines.append(f"FILES BELOW {threshold}% THRESHOLD:")
for path, cov in below_threshold[:10]:
short_path = path.split('/')[-1] if '/' in path else path
lines.append(f" {cov.line_pct:5.1f}% {short_path}")
if len(below_threshold) > 10:
lines.append(f" ... and {len(below_threshold) - 10} more files")
lines.append("")
# Recommendations
lines.append("-" * 60)
lines.append("RECOMMENDATIONS:")
all_recs = (
recs['critical'][:2] + recs['high'][:2] + recs['medium'][:2]
)
for i, rec in enumerate(all_recs[:5], 1):
lines.append(f" {i}. {rec['recommendation']}")
lines.append(f" File: {rec['file']}")
lines.append("")
lines.append("=" * 60)
return '\n'.join(lines)
def generate_html_report(
self,
files: Dict[str, FileCoverage],
summary: CoverageSummary,
analysis: Dict[str, Any],
threshold: int
) -> str:
"""Generate an HTML report"""
stats = analysis['stats']
recs = analysis['recommendations']
html = f"""<!DOCTYPE html>
<html lang="en">
<head>
<meta charset="UTF-8">
<meta name="viewport" content="width=device-width, initial-scale=1.0">
<title>Coverage Analysis Report</title>
<style>
body {{ font-family: -apple-system, BlinkMacSystemFont, 'Segoe UI', Roboto, sans-serif; margin: 40px; }}
h1 {{ color: #333; }}
.summary {{ display: grid; grid-template-columns: repeat(4, 1fr); gap: 20px; margin: 20px 0; }}
.stat {{ background: #f5f5f5; padding: 20px; border-radius: 8px; text-align: center; }}
.stat-value {{ font-size: 2em; font-weight: bold; }}
.pass {{ color: #22c55e; }}
.fail {{ color: #ef4444; }}
.warn {{ color: #f59e0b; }}
table {{ width: 100%; border-collapse: collapse; margin: 20px 0; }}
th, td {{ padding: 12px; text-align: left; border-bottom: 1px solid #ddd; }}
th {{ background: #f5f5f5; }}
.gap-critical {{ background: #fef2f2; }}
.gap-high {{ background: #fffbeb; }}
.progress {{ background: #e5e7eb; border-radius: 4px; height: 8px; }}
.progress-bar {{ height: 100%; border-radius: 4px; }}
</style>
</head>
<body>
<h1>Coverage Analysis Report</h1>
<p>Generated: {datetime.now().strftime('%Y-%m-%d %H:%M:%S')}</p>
<div class="summary">
<div class="stat">
<div class="stat-value {'pass' if stats['overall_statement_pct'] >= threshold else 'fail'}">{stats['overall_statement_pct']:.1f}%</div>
<div>Statements</div>
</div>
<div class="stat">
<div class="stat-value {'pass' if stats['overall_branch_pct'] >= threshold - 5 else 'fail'}">{stats['overall_branch_pct']:.1f}%</div>
<div>Branches</div>
</div>
<div class="stat">
<div class="stat-value {'pass' if stats['overall_function_pct'] >= threshold else 'fail'}">{stats['overall_function_pct']:.1f}%</div>
<div>Functions</div>
</div>
<div class="stat">
<div class="stat-value {'pass' if stats['overall_line_pct'] >= threshold else 'fail'}">{stats['overall_line_pct']:.1f}%</div>
<div>Lines</div>
</div>
</div>
<h2>Threshold Status: <span class="{'pass' if stats['meets_threshold'] else 'fail'}">{'PASS' if stats['meets_threshold'] else 'FAIL'}</span></h2>
<p>Target: {threshold}% | Files Analyzed: {stats['files_analyzed']} | Below Threshold: {stats['files_below_threshold']}</p>
<h2>Coverage Gaps</h2>
<table>
<thead>
<tr>
<th>Severity</th>
<th>File</th>
<th>Issue</th>
<th>Recommendation</th>
</tr>
</thead>
<tbody>
"""
# Add gaps to table
all_gaps = (
[(g, 'critical') for g in recs['critical']] +
[(g, 'high') for g in recs['high']] +
[(g, 'medium') for g in recs['medium'][:5]]
)
for gap, severity in all_gaps[:15]:
row_class = f"gap-{severity}" if severity in ['critical', 'high'] else ""
html += f""" <tr class="{row_class}">
<td>{severity.upper()}</td>
<td>{gap['file'].split('/')[-1]}</td>
<td>{gap['description']}</td>
<td>{gap['recommendation']}</td>
</tr>
"""
html += """ </tbody>
</table>
<h2>File Coverage Details</h2>
<table>
<thead>
<tr>
<th>File</th>
<th>Statements</th>
<th>Branches</th>
<th>Functions</th>
<th>Lines</th>
</tr>
</thead>
<tbody>
"""
# Sort files by line coverage
sorted_files = sorted(files.items(), key=lambda x: x[1].line_pct)
for path, cov in sorted_files[:20]:
short_path = path.split('/')[-1] if '/' in path else path
html += f""" <tr>
<td>{short_path}</td>
<td>{cov.statement_pct:.1f}%</td>
<td>{cov.branch_pct:.1f}%</td>
<td>{cov.function_pct:.1f}%</td>
<td>{cov.line_pct:.1f}%</td>
</tr>
"""
html += """ </tbody>
</table>
</body>
</html>
"""
return html
class CoverageAnalyzerTool:
"""Main tool class"""
def __init__(
self,
coverage_path: str,
threshold: int = 80,
critical_paths: bool = False,
strict: bool = False,
output_format: str = 'text',
output_path: Optional[str] = None,
verbose: bool = False
):
self.coverage_path = Path(coverage_path)
self.threshold = threshold
self.critical_paths = critical_paths
self.strict = strict
self.output_format = output_format
self.output_path = output_path
self.verbose = verbose
def run(self) -> Dict[str, Any]:
"""Run the coverage analysis"""
print(f"Analyzing coverage from: {self.coverage_path}")
# Parse coverage data
parser = CoverageParser(self.verbose)
files, summary = parser.parse(self.coverage_path)
print(f"Found coverage data for {len(files)} files")
# Analyze coverage
analyzer = CoverageAnalyzer(
threshold=self.threshold,
critical_paths=self.critical_paths,
verbose=self.verbose
)
gaps, analysis = analyzer.analyze(files, summary)
# Generate report
reporter = ReportGenerator(self.verbose)
if self.output_format == 'html':
report = reporter.generate_html_report(files, summary, analysis, self.threshold)
else:
report = reporter.generate_text_report(files, summary, analysis, self.threshold)
# Output report
if self.output_path:
with open(self.output_path, 'w') as f:
f.write(report)
print(f"Report written to: {self.output_path}")
else:
print(report)
# Return results
results = {
'status': 'pass' if analysis['stats']['meets_threshold'] else 'fail',
'threshold': self.threshold,
'coverage': {
'statements': analysis['stats']['overall_statement_pct'],
'branches': analysis['stats']['overall_branch_pct'],
'functions': analysis['stats']['overall_function_pct'],
'lines': analysis['stats']['overall_line_pct']
},
'files_analyzed': summary.files_analyzed,
'files_below_threshold': analysis['stats']['files_below_threshold'],
'total_gaps': analysis['stats']['total_gaps'],
'critical_gaps': analysis['stats']['critical_gaps']
}
# Exit with error if strict mode and below threshold
if self.strict and not analysis['stats']['meets_threshold']:
print(f"\nFailed: Coverage {analysis['stats']['overall_line_pct']:.1f}% below threshold {self.threshold}%")
sys.exit(1)
return results
def main():
"""Main entry point"""
parser = argparse.ArgumentParser(
description="Analyze Jest/Istanbul coverage reports and identify gaps",
formatter_class=argparse.RawDescriptionHelpFormatter,
epilog="""
Examples:
# Basic analysis
python coverage_analyzer.py coverage/coverage-final.json
# With threshold enforcement
python coverage_analyzer.py coverage/ --threshold 80 --strict
# Generate HTML report
python coverage_analyzer.py coverage/ --format html --output report.html
# Focus on critical paths
python coverage_analyzer.py coverage/ --critical-paths
"""
)
parser.add_argument(
'coverage',
help='Path to coverage file or directory'
)
parser.add_argument(
'--threshold', '-t',
type=int,
default=80,
help='Coverage threshold percentage (default: 80)'
)
parser.add_argument(
'--strict',
action='store_true',
help='Exit with error if coverage is below threshold'
)
parser.add_argument(
'--critical-paths',
action='store_true',
help='Focus analysis on critical business paths'
)
parser.add_argument(
'--format', '-f',
choices=['text', 'html', 'json'],
default='text',
help='Output format (default: text)'
)
parser.add_argument(
'--output', '-o',
help='Output file path'
)
parser.add_argument(
'--verbose', '-v',
action='store_true',
help='Enable verbose output'
)
parser.add_argument(
'--json',
action='store_true',
help='Output results as JSON (summary only)'
)
args = parser.parse_args()
try:
tool = CoverageAnalyzerTool(
coverage_path=args.coverage,
threshold=args.threshold,
critical_paths=args.critical_paths,
strict=args.strict,
output_format=args.format,
output_path=args.output,
verbose=args.verbose
)
results = tool.run()
if args.json:
print(json.dumps(results, indent=2))
except Exception as e:
print(f"Error: {e}")
if args.verbose:
import traceback
traceback.print_exc()
sys.exit(1)
if __name__ == '__main__':
main()