- AgentHub: 13 files updated with non-engineering examples (content drafts, research, strategy) — engineering stays primary, cross-domain secondary - AgentHub: 7 slash commands, 5 Python scripts, 3 references, 1 agent, dry_run.py validation (57 checks) - Marketplace: agenthub entry added with cross-domain keywords, engineering POWERFUL updated (25→30), product (12→13), counts synced across all configs - SEO: generate-docs.py now produces keyword-rich <title> tags and meta descriptions using SKILL.md frontmatter — "Claude Code Skills" in site_name propagates to all 276 HTML pages - SEO: per-domain title suffixes (Agent Skill for Codex & OpenClaw, etc.), slug-as-title cleanup, domain label stripping from titles - Broken links: 141→0 warnings — new rewrite_skill_internal_links() converts references/, scripts/, assets/ links to GitHub source URLs; skills/index.md phantom slugs fixed (6 marketing, 7 RA/QM) - Counts synced: 204 skills, 266 tools, 382 refs, 16 agents, 17 commands, 21 plugins — consistent across CLAUDE.md, README.md, docs/index.md, marketplace.json, getting-started.md, mkdocs.yml - Platform sync: Codex 163 skills, Gemini 246 items, OpenClaw compatible Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
2.6 KiB
2.6 KiB
title, description
| title | description |
|---|---|
| Interview System Designer — Agent Skill for Codex & OpenClaw | This skill should be used when the user asks to 'design interview processes', 'create hiring pipelines', 'calibrate interview loops', 'generate. Agent skill for Claude Code, Codex CLI, Gemini CLI, OpenClaw. |
Interview System Designer
:material-rocket-launch: Engineering - POWERFUL
:material-identifier: `interview-system-designer`
:material-github: Source
Install:
claude /plugin install engineering-advanced-skills
Comprehensive interview loop planning and calibration support for role-based hiring systems.
Overview
Use this skill to create structured interview loops, standardize question quality, and keep hiring signal consistent across interviewers.
Core Capabilities
- Interview loop planning by role and level
- Round-by-round focus and timing recommendations
- Suggested question sets by round type
- Framework support for scoring and calibration
- Bias-reduction and process consistency guidance
Quick Start
# Generate a loop plan for a role and level
python3 scripts/interview_planner.py --role "Senior Software Engineer" --level senior
# JSON output for integration with internal tooling
python3 scripts/interview_planner.py --role "Product Manager" --level mid --json
Recommended Workflow
- Run
scripts/interview_planner.pyto generate a baseline loop. - Align rounds to role-specific competencies.
- Validate scoring rubric consistency with interview panel leads.
- Review for bias controls before rollout.
- Recalibrate quarterly using hiring outcome data.
References
references/interview-frameworks.mdreferences/bias_mitigation_checklist.mdreferences/competency_matrix_templates.mdreferences/debrief_facilitation_guide.md
Common Pitfalls
- Overweighting one round while ignoring other competency signals
- Using unstructured interviews without standardized scoring
- Skipping calibration sessions for interviewers
- Changing hiring bar without documenting rationale
Best Practices
- Keep round objectives explicit and non-overlapping.
- Require evidence for each score recommendation.
- Use the same baseline rubric across comparable roles.
- Revisit loop design based on quality-of-hire outcomes.