* feat: Skill Authoring Standard + Marketing Expansion plans
SKILL-AUTHORING-STANDARD.md — the DNA of every skill in this repo:
10 universal patterns codified from C-Suite innovations + Corey Haines' marketingskills patterns:
1. Context-First: check domain context, ask only for gaps
2. Practitioner Voice: expert persona, goal-oriented, not textbook
3. Multi-Mode Workflows: build from scratch / optimize existing / situation-specific
4. Related Skills Navigation: when to use, when NOT to, bidirectional
5. Reference Separation: SKILL.md lean (≤10KB), refs deep
6. Proactive Triggers: surface issues without being asked
7. Output Artifacts: request → specific deliverable mapping
8. Quality Loop: self-verify, confidence tagging
9. Communication Standard: bottom line first, structured output
10. Python Tools: stdlib-only, CLI-first, JSON output, sample data
Marketing expansion plans for 40-skill marketing division build.
* feat: marketing foundation — context + ops router + authoring standard
marketing-context/: Foundation skill every marketing skill reads first
- SKILL.md: 3 modes (auto-draft, guided interview, update)
- templates/marketing-context-template.md: 14 sections covering
product, audience, personas, pain points, competitive landscape,
differentiation, objections, switching dynamics, customer language
(verbatim), brand voice, style guide, proof points, SEO context, goals
- scripts/context_validator.py: Scores completeness 0-100, section-by-section
marketing-ops/: Central router for 40-skill marketing ecosystem
- Full routing matrix: 7 pods + cross-domain routing to 6 skills in
business-growth, product-team, engineering-team, c-level-advisor
- Campaign orchestration sequences (launch, content, CRO sprint)
- Quality gate matching C-Suite standard
- scripts/campaign_tracker.py: Campaign status tracking with progress,
overdue detection, pod coverage, blocker identification
SKILL-AUTHORING-STANDARD.md: Universal DNA for all skills
- 10 patterns: context-first, practitioner voice, multi-mode workflows,
related skills navigation, reference separation, proactive triggers,
output artifacts, quality loop, communication standard, python tools
- Quality checklist for skill completion verification
- Domain context file mapping for all 5 domains
* feat: import 20 workspace marketing skills + standard sections
Imported 20 marketing skills from OpenClaw workspace into repo:
Content Pod (5):
content-strategy, copywriting, copy-editing, social-content, marketing-ideas
SEO Pod (2):
seo-audit (+ references enriched by subagent), programmatic-seo (+ refs)
CRO Pod (5):
page-cro, form-cro, signup-flow-cro, onboarding-cro, popup-cro, paywall-upgrade-cro
Channels Pod (2):
email-sequence, paid-ads
Growth + Intel + GTM (5):
ab-test-setup, competitor-alternatives, marketing-psychology, launch-strategy, brand-guidelines
All 29 skills now have standard sections per SKILL-AUTHORING-STANDARD.md:
✅ Proactive Triggers (4-5 per skill)
✅ Output Artifacts table
✅ Communication standard reference
✅ Related Skills with WHEN/NOT disambiguation
Subagents enriched 8 skills with additional reference docs:
seo-audit, programmatic-seo, page-cro, form-cro,
onboarding-cro, popup-cro, paywall-upgrade-cro, email-sequence
43 files, 10,566 lines added.
* feat: build 13 new marketing skills + social-media-manager upgrade
All skills are 100% original work — inspired by industry best practices,
written from scratch in our own voice following SKILL-AUTHORING-STANDARD.md.
NEW Content Pod (2):
content-production — full research→draft→optimize pipeline, content_scorer.py
content-humanizer — AI pattern detection + voice injection, humanizer_scorer.py
NEW SEO Pod (3):
ai-seo — AI search optimization (AEO/GEO/LLMO), entirely new category
schema-markup — JSON-LD structured data, schema_validator.py
site-architecture — URL structure + internal linking, sitemap_analyzer.py
NEW Channels Pod (2):
cold-email — B2B outreach (distinct from email-sequence lifecycle)
ad-creative — bulk ad generation + platform specs, ad_copy_validator.py
NEW Growth Pod (3):
churn-prevention — cancel flows + save offers + dunning, churn_impact_calculator.py
referral-program — referral + affiliate programs
free-tool-strategy — engineering as marketing
NEW Intelligence Pod (1):
analytics-tracking — GA4/GTM setup + event taxonomy, tracking_plan_generator.py
NEW Sales Pod (1):
pricing-strategy — pricing, packaging, monetization
UPGRADED:
social-media-analyzer → social-media-manager (strategy, calendar, community)
Totals: 42 skills, 27 Python scripts, 60 reference docs, 163 files, 43,265 lines
* feat: update index, marketplace, README for 42 marketing skills
- skills-index.json: 89 → 124 skills (42 marketing entries)
- marketplace.json: marketing-skills v2.0.0 (42 skills, 27 tools)
- README.md: badge 134 → 169, marketing row updated
- prompt-engineer-toolkit: added YAML frontmatter
- Removed build logs from repo
- Parity check: 42/42 passed (YAML + Related + Proactive + Output + Communication)
* fix: merge content-creator into content-production, split marketing-psychology
Quality audit fixes:
1. content-creator → DEPRECATED redirect
- Scripts (brand_voice_analyzer.py, seo_optimizer.py) moved to content-production
- SKILL.md replaced with redirect to content-production + content-strategy
- Eliminates duplicate routing confusion
2. marketing-psychology → 24KB split to 6.8KB + reference
- 70+ mental models moved to references/mental-models-catalog.md (397 lines)
- SKILL.md now lean: categories overview, most-used models, quick reference
- Saves ~4,300 tokens per invocation
* feat: add plugin configs, Codex/OpenClaw compatibility, ClawHub packaging
- marketing-skill/SKILL.md: ClawHub-compatible root with Quick Start for Claude Code, Codex CLI, OpenClaw
- marketing-skill/CLAUDE.md: Agent instructions (routing, context, anti-patterns)
- marketing-skill/.codex/instructions.md: Codex CLI skill routing
- .claude-plugin/marketplace.json: deduplicated, marketing-skills v2.0.0
- .codex/skills-index.json: content-creator marked deprecated, psychology updated
- Total: 42 skills, 27 Python tools, 60 references, 18 plugins
* feat: add 16 Python tools to knowledge-only skills
Enriched 12 previously tool-less skills with practical Python scripts:
- seo-audit/seo_checker.py — HTML on-page SEO analysis (0-100)
- copywriting/headline_scorer.py — headline quality scoring (0-100)
- copy-editing/readability_scorer.py — Flesch + passive + filler detection
- content-strategy/topic_cluster_mapper.py — keyword clustering
- page-cro/conversion_audit.py — HTML CRO signal analysis (0-100)
- paid-ads/roas_calculator.py — ROAS/CPA/CPL calculator
- email-sequence/sequence_analyzer.py — email sequence scoring (0-100)
- form-cro/form_field_analyzer.py — form field CRO audit (0-100)
- onboarding-cro/activation_funnel_analyzer.py — funnel drop-off analysis
- programmatic-seo/url_pattern_generator.py — URL pattern planning
- ab-test-setup/sample_size_calculator.py — statistical sample sizing
- signup-flow-cro/funnel_drop_analyzer.py — signup funnel analysis
- launch-strategy/launch_readiness_scorer.py — launch checklist scoring
- competitor-alternatives/comparison_matrix_builder.py — feature comparison
- social-media-manager/social_calendar_generator.py — content calendar
- readability_scorer.py — fixed demo mode for non-TTY execution
All 43/43 scripts pass execution. All stdlib-only, zero pip installs.
Total: 42 skills, 43 Python tools, 60+ reference docs.
* feat: add 3 more Python tools + improve 6 existing scripts
New tools from build agent:
- email-sequence/scripts/sequence_analyzer.py — email sequence scoring (91/100 demo)
- paid-ads/scripts/roas_calculator.py — ROAS/CPA/CPL/break-even calculator
- competitor-alternatives/scripts/comparison_matrix_builder.py — feature matrix
Improved scripts (better demo modes, fuller analysis):
- seo_checker.py, headline_scorer.py, readability_scorer.py,
conversion_audit.py, topic_cluster_mapper.py, launch_readiness_scorer.py
Total: 42 skills, 47 Python tools, all passing.
* fix: remove duplicate scripts from deprecated content-creator
Scripts already live in content-production/scripts/. The content-creator
directory is now a pure redirect (SKILL.md only + legacy assets/refs).
* fix: scope VirusTotal scan to executable files only
Skip scanning .md, .py, .json, .yml — they're plain text files
that VirusTotal can't meaningfully analyze. This prevents 429 rate
limit errors on PRs with many text file changes (like 42 marketing skills).
Scan still covers: .js, .ts, .sh, .mjs, .cjs, .exe, .dll, .so, .bin, .wasm
---------
Co-authored-by: Leo <leo@openclaw.ai>
355 lines
13 KiB
Python
Executable File
355 lines
13 KiB
Python
Executable File
#!/usr/bin/env python3
|
|
"""
|
|
sequence_analyzer.py — Email sequence quality analyzer
|
|
Usage:
|
|
python3 sequence_analyzer.py --file sequence.json
|
|
python3 sequence_analyzer.py --json
|
|
python3 sequence_analyzer.py # demo mode
|
|
|
|
Input JSON format:
|
|
[
|
|
{"subject": "...", "body": "...", "delay_days": 0},
|
|
{"subject": "...", "body": "...", "delay_days": 2},
|
|
...
|
|
]
|
|
"""
|
|
|
|
import argparse
|
|
import json
|
|
import re
|
|
import sys
|
|
|
|
|
|
# ---------------------------------------------------------------------------
|
|
# Word/pattern lists
|
|
# ---------------------------------------------------------------------------
|
|
|
|
SPAM_TRIGGER_WORDS = [
|
|
"free", "guarantee", "guaranteed", "winner", "won", "prize",
|
|
"congratulations", "cash", "earn money", "make money", "extra income",
|
|
"100% free", "no cost", "risk free", "act now", "limited time",
|
|
"click here", "buy now", "order now", "get it now",
|
|
"as seen on", "dear friend", "you have been selected",
|
|
"this isn't spam", "not spam", "no credit card required",
|
|
"special promotion", "special offer", "amazing offer",
|
|
"!!!", "!!!", "$$$", "£££",
|
|
"increase your", "increase sales", "double your",
|
|
"lose weight", "weight loss", "diet", "viagra", "casino",
|
|
]
|
|
|
|
CTA_PATTERNS = re.compile(
|
|
r"\b(click|tap|reply|download|sign up|register|buy|purchase|get started|"
|
|
r"learn more|read more|visit|go to|check out|schedule|book|claim|try|"
|
|
r"subscribe|join|start|access|watch|see|grab|discover)\b",
|
|
re.IGNORECASE,
|
|
)
|
|
|
|
PERSONALIZATION_TOKENS = re.compile(
|
|
r"\{\{?\s*\w+\s*\}?\}|%\w+%|\[FIRST_NAME\]|\[NAME\]|\[COMPANY\]|\[FIRSTNAME\]",
|
|
re.IGNORECASE,
|
|
)
|
|
|
|
|
|
# ---------------------------------------------------------------------------
|
|
# Per-email analysis
|
|
# ---------------------------------------------------------------------------
|
|
|
|
def analyze_email(email: dict, index: int) -> dict:
|
|
subject = email.get("subject", "")
|
|
body = email.get("body", "")
|
|
delay = email.get("delay_days", 0)
|
|
|
|
# Subject analysis
|
|
subject_len = len(subject)
|
|
subject_word_count = len(subject.split())
|
|
subject_ok = 30 <= subject_len <= 60
|
|
subject_has_number = bool(re.search(r"\d", subject))
|
|
subject_question = subject.strip().endswith("?")
|
|
subject_all_caps = subject == subject.upper() and len(subject) > 3
|
|
|
|
# Body analysis
|
|
body_words = re.findall(r"\b\w+\b", body)
|
|
body_word_count = len(body_words)
|
|
|
|
# CTA detection
|
|
cta_matches = CTA_PATTERNS.findall(body)
|
|
has_cta = len(cta_matches) > 0
|
|
|
|
# Personalization tokens
|
|
tokens_in_subject = PERSONALIZATION_TOKENS.findall(subject)
|
|
tokens_in_body = PERSONALIZATION_TOKENS.findall(body)
|
|
total_tokens = len(tokens_in_subject) + len(tokens_in_body)
|
|
|
|
# Spam triggers
|
|
combined = (subject + " " + body).lower()
|
|
spam_found = [w for w in SPAM_TRIGGER_WORDS if w.lower() in combined]
|
|
|
|
# Spam score (0-100, higher = more spammy)
|
|
spam_score = min(100, len(spam_found) * 10)
|
|
|
|
return {
|
|
"email_index": index + 1,
|
|
"delay_days": delay,
|
|
"subject": {
|
|
"text": subject,
|
|
"length": subject_len,
|
|
"word_count": subject_word_count,
|
|
"length_ok": subject_ok,
|
|
"has_number": subject_has_number,
|
|
"is_question": subject_question,
|
|
"all_caps_warning": subject_all_caps,
|
|
"personalized": len(tokens_in_subject) > 0,
|
|
},
|
|
"body": {
|
|
"word_count": body_word_count,
|
|
"length_verdict": _body_length_verdict(body_word_count),
|
|
"has_cta": has_cta,
|
|
"cta_phrases": list(set(cta_matches))[:5],
|
|
"personalization_tokens": total_tokens,
|
|
},
|
|
"spam": {
|
|
"trigger_words_found": spam_found[:8],
|
|
"trigger_count": len(spam_found),
|
|
"spam_risk_score": spam_score,
|
|
"risk_level": "High" if spam_score >= 40 else "Medium" if spam_score >= 20 else "Low",
|
|
},
|
|
}
|
|
|
|
|
|
def _body_length_verdict(word_count: int) -> str:
|
|
if word_count < 50:
|
|
return "Too short (<50 words)"
|
|
if word_count <= 150:
|
|
return "Short/punchy — good for re-engagement"
|
|
if word_count <= 300:
|
|
return "Optimal (150-300 words)"
|
|
if word_count <= 500:
|
|
return "Long — ensure high value throughout"
|
|
return "Very long (500+ words) — consider trimming"
|
|
|
|
|
|
# ---------------------------------------------------------------------------
|
|
# Sequence-level analysis
|
|
# ---------------------------------------------------------------------------
|
|
|
|
def analyze_pacing(emails: list) -> dict:
|
|
if len(emails) <= 1:
|
|
return {"note": "Single email — no pacing to analyze"}
|
|
|
|
delays = [e.get("delay_days", 0) for e in emails]
|
|
gaps = [delays[i] - delays[i - 1] for i in range(1, len(delays))]
|
|
|
|
issues = []
|
|
for i, gap in enumerate(gaps):
|
|
if gap <= 0:
|
|
issues.append(f"Email {i+2}: same-day or before previous — check delay_days")
|
|
elif gap == 1:
|
|
issues.append(f"Email {i+2}: only 1-day gap — may feel aggressive")
|
|
elif gap > 14:
|
|
issues.append(f"Email {i+2}: {gap}-day gap — momentum may drop")
|
|
|
|
# Assess overall cadence
|
|
avg_gap = sum(gaps) / len(gaps) if gaps else 0
|
|
if avg_gap <= 2:
|
|
cadence = "Aggressive (avg <2 days)"
|
|
elif avg_gap <= 5:
|
|
cadence = "High-frequency (avg 2-5 days)"
|
|
elif avg_gap <= 10:
|
|
cadence = "Standard (avg 5-10 days)"
|
|
else:
|
|
cadence = "Low-frequency (avg 10+ days)"
|
|
|
|
return {
|
|
"email_count": len(emails),
|
|
"total_duration_days": max(delays) - min(delays),
|
|
"avg_gap_days": round(avg_gap, 1),
|
|
"cadence_type": cadence,
|
|
"gaps": gaps,
|
|
"issues": issues,
|
|
}
|
|
|
|
|
|
# ---------------------------------------------------------------------------
|
|
# Scoring
|
|
# ---------------------------------------------------------------------------
|
|
|
|
def compute_sequence_score(email_analyses: list, pacing: dict) -> dict:
|
|
if not email_analyses:
|
|
return {"overall": 0}
|
|
|
|
# Subject score: avg subject length compliance
|
|
subject_ok_count = sum(1 for e in email_analyses if e["subject"]["length_ok"])
|
|
subject_score = round(subject_ok_count / len(email_analyses) * 100)
|
|
|
|
# CTA score: % of emails with CTA
|
|
cta_count = sum(1 for e in email_analyses if e["body"]["has_cta"])
|
|
cta_score = round(cta_count / len(email_analyses) * 100)
|
|
|
|
# Personalization score
|
|
personalized_count = sum(1 for e in email_analyses if e["body"]["personalization_tokens"] > 0)
|
|
personalization_score = round(personalized_count / len(email_analyses) * 100)
|
|
|
|
# Spam score (inverted — low spam = high score)
|
|
avg_spam = sum(e["spam"]["spam_risk_score"] for e in email_analyses) / len(email_analyses)
|
|
spam_score = max(0, 100 - int(avg_spam))
|
|
|
|
# Pacing score
|
|
pacing_issues = len(pacing.get("issues", []))
|
|
pacing_score = max(0, 100 - pacing_issues * 20)
|
|
|
|
# Body length score
|
|
length_ok_count = sum(
|
|
1 for e in email_analyses
|
|
if "Optimal" in e["body"]["length_verdict"] or "punchy" in e["body"]["length_verdict"]
|
|
)
|
|
length_score = round(length_ok_count / len(email_analyses) * 100)
|
|
|
|
weights = {
|
|
"subject_quality": 0.20,
|
|
"cta_presence": 0.20,
|
|
"spam_safety": 0.25,
|
|
"personalization": 0.15,
|
|
"pacing": 0.10,
|
|
"body_length": 0.10,
|
|
}
|
|
scores = {
|
|
"subject_quality": subject_score,
|
|
"cta_presence": cta_score,
|
|
"spam_safety": spam_score,
|
|
"personalization": personalization_score,
|
|
"pacing": pacing_score,
|
|
"body_length": length_score,
|
|
}
|
|
overall = round(sum(scores[k] * weights[k] for k in weights))
|
|
grade = "A" if overall >= 85 else "B" if overall >= 70 else "C" if overall >= 55 else "D" if overall >= 40 else "F"
|
|
|
|
return {
|
|
"overall": overall,
|
|
"grade": grade,
|
|
"breakdown": {k: {"score": v, "weight": f"{int(weights[k]*100)}%"} for k, v in scores.items()},
|
|
}
|
|
|
|
|
|
# ---------------------------------------------------------------------------
|
|
# Demo data
|
|
# ---------------------------------------------------------------------------
|
|
|
|
DEMO_SEQUENCE = [
|
|
{
|
|
"subject": "{{first_name}}, your free marketing audit is ready",
|
|
"body": "Hi {{first_name}},\n\nWe analyzed 500 campaigns like yours and found three quick wins that could double your ROAS in 30 days.\n\nI've put together a custom audit for {{company}}. It's free and takes 10 minutes to review.\n\n→ Click here to see your results: [LINK]\n\nBest,\nSarah",
|
|
"delay_days": 0,
|
|
},
|
|
{
|
|
"subject": "Did you see this, {{first_name}}?",
|
|
"body": "Quick follow-up.\n\nMost marketers we talk to are sitting on 2-3 easy optimizations that could add 20-40% more revenue from the same ad spend.\n\nHere's the #1 thing we see: landing pages that don't match the ad promise.\n\nWorth 5 minutes? → [Review your audit]\n\nSarah",
|
|
"delay_days": 3,
|
|
},
|
|
{
|
|
"subject": "The $50,000 mistake (and how to avoid it)",
|
|
"body": "True story.\n\nOne of our clients was spending $8,500/month on Google Ads with a 1.8x ROAS. Technically above break-even, but barely.\n\nWe found that 60% of their budget was going to one keyword that had zero purchase intent.\n\nAfter fixing it: same spend, 4.2x ROAS.\n\nThat's the kind of thing our audit catches. Have you looked at yours yet?\n\n→ [Open your free audit]\n\nSarah\n\nP.S. This offer expires Friday.",
|
|
"delay_days": 5,
|
|
},
|
|
{
|
|
"subject": "Last call — your audit expires tonight",
|
|
"body": "{{first_name}}, this is the last reminder.\n\nYour personalized audit expires at midnight tonight.\n\nIf growing your ROAS is a priority this quarter, take 10 minutes now.\n\n→ [Claim your audit before it expires]\n\nSarah",
|
|
"delay_days": 7,
|
|
},
|
|
{
|
|
"subject": "New case study: {{company}}-style win",
|
|
"body": "Since you didn't grab the audit, I wanted to send you something valuable anyway.\n\nHere's a 3-minute case study showing how we helped a B2B SaaS company go from 1.9x to 5.4x ROAS in 45 days.\n\nNo audit required — just solid tactics you can steal.\n\n→ [Read the case study]\n\nHope it helps,\nSarah",
|
|
"delay_days": 14,
|
|
},
|
|
]
|
|
|
|
|
|
# ---------------------------------------------------------------------------
|
|
# Main
|
|
# ---------------------------------------------------------------------------
|
|
|
|
def main():
|
|
parser = argparse.ArgumentParser(
|
|
description="Email sequence analyzer — scores sequence quality 0-100."
|
|
)
|
|
parser.add_argument("--file", help="JSON file with email sequence array")
|
|
parser.add_argument("--json", action="store_true", help="Output as JSON")
|
|
args = parser.parse_args()
|
|
|
|
if args.file:
|
|
with open(args.file, "r", encoding="utf-8") as f:
|
|
emails = json.load(f)
|
|
else:
|
|
emails = DEMO_SEQUENCE
|
|
if not args.json:
|
|
print("No input provided — running in demo mode (5-email nurture sequence).\n")
|
|
|
|
email_analyses = [analyze_email(e, i) for i, e in enumerate(emails)]
|
|
pacing = analyze_pacing(emails)
|
|
scoring = compute_sequence_score(email_analyses, pacing)
|
|
|
|
if args.json:
|
|
output = {
|
|
"sequence_score": scoring,
|
|
"pacing": pacing,
|
|
"emails": email_analyses,
|
|
}
|
|
print(json.dumps(output, indent=2))
|
|
return
|
|
|
|
# Human-readable
|
|
overall = scoring["overall"]
|
|
grade = scoring["grade"]
|
|
|
|
print("=" * 64)
|
|
print(f" EMAIL SEQUENCE ANALYSIS Score: {overall}/100 Grade: {grade}")
|
|
print("=" * 64)
|
|
|
|
# Pacing summary
|
|
print(f"\n 📅 SEQUENCE PACING")
|
|
print(f" Emails: {pacing['email_count']}")
|
|
print(f" Duration: {pacing.get('total_duration_days', 0)} days")
|
|
print(f" Avg gap: {pacing.get('avg_gap_days', 0)} days")
|
|
print(f" Cadence: {pacing.get('cadence_type', 'N/A')}")
|
|
if pacing.get("issues"):
|
|
for issue in pacing["issues"]:
|
|
print(f" ⚠️ {issue}")
|
|
|
|
print(f"\n 📧 PER-EMAIL BREAKDOWN")
|
|
print(f" {'#':<3} {'Subject':<40} {'Words':<6} {'CTA':<4} {'Tokens':<7} {'Spam'}")
|
|
print(" " + "─" * 60)
|
|
|
|
for e in email_analyses:
|
|
subj = e["subject"]["text"][:38]
|
|
if not e["subject"]["length_ok"]:
|
|
subj += "⚠️"
|
|
words = e["body"]["word_count"]
|
|
cta = "✅" if e["body"]["has_cta"] else "❌"
|
|
tokens = e["body"]["personalization_tokens"]
|
|
spam_lvl = e["spam"]["risk_level"]
|
|
spam_icon = "✅" if spam_lvl == "Low" else ("⚠️ " if spam_lvl == "Medium" else "❌")
|
|
spam_str = f"{spam_icon}{spam_lvl}"
|
|
print(f" {e['email_index']:<3} {subj:<40} {words:<6} {cta:<4} {tokens:<7} {spam_str}")
|
|
|
|
if any(e["spam"]["trigger_words_found"] for e in email_analyses):
|
|
print(f"\n ⚠️ SPAM TRIGGER WORDS DETECTED")
|
|
for e in email_analyses:
|
|
if e["spam"]["trigger_words_found"]:
|
|
triggers = ", ".join(e["spam"]["trigger_words_found"])
|
|
print(f" Email {e['email_index']}: {triggers}")
|
|
|
|
print(f"\n SCORE BREAKDOWN")
|
|
for k, v in scoring["breakdown"].items():
|
|
label = k.replace("_", " ").title()
|
|
bar_len = round(v["score"] / 10)
|
|
bar = "█" * bar_len + "░" * (10 - bar_len)
|
|
print(f" {label:<22} [{bar}] {v['score']:>3}/100 (weight {v['weight']})")
|
|
|
|
print()
|
|
print("=" * 64)
|
|
print(f" Overall: {overall}/100 Grade: {grade}")
|
|
print("=" * 64)
|
|
|
|
|
|
if __name__ == "__main__":
|
|
main()
|