bump-codex-defaults

Bump codex defaults to gpt-5.5 worker (keep gpt-5.4-mini meta)

Metadata

Statusdone
Assignedagent-1030
Agent identityf51439356729d112a6c404803d88015d5b44832c6c584c62b96732b63c2b0c7e
Created2026-04-29T00:10:51.455197210+00:00
Started2026-04-29T00:11:10.794901232+00:00
Completed2026-04-29T00:26:14.764889671+00:00
Tagsfix,codex,config, eval-scheduled
Eval score0.91
└ blocking impact0.90
└ completeness0.95
└ coordination overhead0.90
└ correctness0.95
└ downstream usability0.90
└ efficiency0.85
└ intent fidelity0.88
└ style adherence0.95

Description

Description

Update the codex defaults to use gpt-5.5 (newest frontier model, / per MTok) as the worker default instead of gpt-5.4 ($2.50/$15). User preference: prefer newest capability for workers; the 2x cost difference is acceptable given codex:gpt-5.5 is still 3x cheaper than claude:opus per MTok. Meta-tasks (eval/flip/assign) stay on gpt-5.4-mini (best mini tier — no gpt-5.5-mini exists).

Scope

  1. wg init --route codex-cli should write:

    • [agent].model = "codex:gpt-5.5" (was: gpt-5.4)
    • [dispatcher].model = "codex:gpt-5.5" (was: gpt-5.4)
    • [models.default].model = "codex:gpt-5.5" (was: gpt-5.4)
    • [models.evaluator] / [models.assigner] stay on codex:gpt-5.4-mini (unchanged)
    • [tiers] section: fast=gpt-5.4-mini (unchanged), standard=gpt-5.5 (was gpt-5.4), premium=gpt-5.5 (unchanged) — OR keep standard=gpt-5.4 and just change [agent]/[dispatcher]. Pick the consistent option.
  2. The codex starter profile (written by wg profile init-starters) should match — worker on gpt-5.5, meta on gpt-5.4-mini. The current starter description says "gpt-5.5 worker, gpt-5.4-mini for agency" which is what we want; verify it matches the actual file content.

  3. Add a [models.flip] entry (currently missing in the codex-cli route output) so FLIP scoring also explicitly uses gpt-5.4-mini instead of falling through to a default that might silently be claude:haiku.

Validation

  • wg init --route codex-cli --dry-run output shows codex:gpt-5.5 in [agent].model and [dispatcher].model
  • wg init --route codex-cli --dry-run output includes a [models.flip] section pointing at codex:gpt-5.4-mini
  • wg profile init-starters followed by wg profile show codex shows gpt-5.5 as worker model
  • Live smoke: in a fresh tmpdir, wg init --route codex-cli then spawn a tiny task and confirm worker uses gpt-5.5 (wg agents or WG_MODEL env var)
  • cargo build + cargo test pass with no regressions
  • cargo install --path . was run before claiming done

Depends on

Required by

Log