From a3ba6f825680808cd1b63143b9a7d1a7cb319ed3 Mon Sep 17 00:00:00 2001 From: Juan Michelini Date: Wed, 4 Mar 2026 18:40:02 -0300 Subject: [PATCH 1/2] Update model configuration for GPT-5.3 Codex --- .github/run-eval/resolve_model_config.py | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/.github/run-eval/resolve_model_config.py b/.github/run-eval/resolve_model_config.py index c21ecc8888..3e60695b66 100755 --- a/.github/run-eval/resolve_model_config.py +++ b/.github/run-eval/resolve_model_config.py @@ -127,10 +127,10 @@ "display_name": "GPT-5.2 Codex", "llm_config": {"model": "litellm_proxy/gpt-5.2-codex"}, }, - "gpt-5-3-codex": { - "id": "gpt-5-3-codex", + "gpt-5.3-codex": { + "id": "gpt-5.3-codex", "display_name": "GPT-5.3 Codex", - "llm_config": {"model": "litellm_proxy/gpt-5-3-codex"}, + "llm_config": {"model": "litellm_proxy/gpt-5.3-codex"}, }, "gpt-5.2-high-reasoning": { "id": "gpt-5.2-high-reasoning", From a77b25a9008edffaab86375d58529b5f3265fcdf Mon Sep 17 00:00:00 2001 From: openhands Date: Wed, 4 Mar 2026 22:10:32 +0000 Subject: [PATCH 2/2] Update test to use corrected gpt-5.3-codex model key Fix test_gpt_5_3_codex_config to reference the corrected model key 'gpt-5.3-codex' (with dots) instead of the old 'gpt-5-3-codex' (with hyphens) to match the fix in resolve_model_config.py. Co-authored-by: openhands --- tests/github_workflows/test_resolve_model_config.py | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/tests/github_workflows/test_resolve_model_config.py b/tests/github_workflows/test_resolve_model_config.py index 2ed496df68..b59fa0a7b4 100644 --- a/tests/github_workflows/test_resolve_model_config.py +++ b/tests/github_workflows/test_resolve_model_config.py @@ -250,12 +250,12 @@ def test_gpt_oss_20b_config(): def test_gpt_5_3_codex_config(): - """Test that gpt-5-3-codex has correct configuration.""" - model = MODELS["gpt-5-3-codex"] + """Test that gpt-5.3-codex has correct configuration.""" + model = MODELS["gpt-5.3-codex"] - assert model["id"] == "gpt-5-3-codex" + assert model["id"] == "gpt-5.3-codex" assert model["display_name"] == "GPT-5.3 Codex" - assert model["llm_config"]["model"] == "litellm_proxy/gpt-5-3-codex" + assert model["llm_config"]["model"] == "litellm_proxy/gpt-5.3-codex" def test_glm_5_config():