Demo repository for showcasing CodexOpt on intentionally messy instruction assets.
AGENTS.mdwith duplicate and conflicting guidanceSKILL.mdexamples:- missing frontmatter
- verbose/redundant text
- duplicated lines
tasks.mdwith 5 evaluation tasks- Tiny Python package under
src/codexopt_demo - GEPA local/cloud setup guide:
docs/gepa-local-and-cloud.md
uv lock
uv sync --extra dev
uv run --no-sync pytest -q
uv run --no-sync ruff check src testsFrom this repo root:
codexopt init
codexopt scan
codexopt benchmark
codexopt optimize agents --file AGENTS.md
codexopt optimize skills --glob ".codex/skills/**/SKILL.md"
codexopt apply --kind skills --dry-run
codexopt report --output codexopt-report.mdUse this example file:
codexopt.gepa.example.yaml
cp codexopt.gepa.example.yaml codexopt.yamlEdit codexopt.yaml:
optimization:
engine: "gepa"
max_metric_calls: 120
reflection_model: "your-provider/your-reflection-model"codexopt optimize agents --config codexopt.yaml
codexopt optimize skills --config codexopt.yamlcodexopt optimize skills \
--engine gepa \
--reflection-model your-provider/your-reflection-model \
--max-metric-calls 200Current CodexOpt exposes GEPA tuning via max_metric_calls and reflection_model.
A direct iterations field is not exposed yet; use max_metric_calls as the primary search-budget control.
For step-by-step local and cloud GEPA setup (including low-budget runs), see:
docs/gepa-local-and-cloud.md