Files
claude-code-skills-reference/product-analysis/references/codex_patterns.md
2026-02-23 16:16:58 +08:00

2.7 KiB

Codex CLI Integration Patterns

How to use OpenAI Codex CLI for cross-model parallel analysis.

Basic Invocation

codex -m o4-mini \
  -c model_reasoning_effort="high" \
  --full-auto \
  "Your analysis prompt here"

Flag Reference

Flag Purpose Values
-m Model selection o4-mini (fast), gpt-5.3-codex-spark (deep)
-c model_reasoning_effort Reasoning depth low, medium, high, xhigh
-c model_reasoning_summary_format Summary format experimental (structured output)
--full-auto Skip all approval prompts (no value)
--dangerously-bypass-approvals-and-sandbox Legacy full-auto flag (no value, older versions)

Fast Scan (quick validation)

codex -m o4-mini \
  -c model_reasoning_effort="medium" \
  --full-auto \
  "prompt"

Deep Analysis (thorough investigation)

codex -m o4-mini \
  -c model_reasoning_effort="xhigh" \
  -c model_reasoning_summary_format="experimental" \
  --full-auto \
  "prompt"

Parallel Execution Pattern

Launch multiple Codex analyses in background using Bash tool with run_in_background: true:

# Dimension 1: Frontend
codex -m o4-mini -c model_reasoning_effort="high" --full-auto \
  "Analyze frontend navigation: count interactive elements, find duplicate entry points, assess cognitive load for new users. Give file paths and counts."

# Dimension 2: User Journey
codex -m o4-mini -c model_reasoning_effort="high" --full-auto \
  "Analyze new user experience: what does empty state show? How many steps to first action? Count clickable elements competing for attention. Give file paths."

# Dimension 3: Backend API
codex -m o4-mini -c model_reasoning_effort="high" --full-auto \
  "List all API endpoints. Identify unused endpoints with no frontend consumer. Check error handling consistency. Give router file paths."

Output Handling

Codex outputs to stdout. When run in background:

  1. Use Bash run_in_background: true to launch
  2. Use TaskOutput to retrieve results when done
  3. Parse the text output for findings

Cross-Model Value

The primary value of Codex in this workflow is independent perspective:

  • Different training data may surface different patterns
  • Different reasoning approach may catch what Claude misses
  • Agreement across models = high confidence
  • Disagreement = worth investigating manually

Limitations

  • Codex CLI must be installed and configured (codex command available)
  • Requires OpenAI API key configured
  • No MCP server access (only filesystem tools)
  • Output is unstructured text (needs parsing)
  • Rate limits apply per OpenAI account