From 25e436a4aa96c62275680fc0e627ce91c7772663 Mon Sep 17 00:00:00 2001 From: YeonGyu-Kim Date: Fri, 6 Feb 2026 15:12:41 +0900 Subject: [PATCH] fix: update snapshots and remove duplicate key in switcher for model version update --- .../__snapshots__/model-fallback.test.ts.snap | 124 +++++++++--------- src/hooks/think-mode/switcher.ts | 1 - 2 files changed, 62 insertions(+), 63 deletions(-) diff --git a/src/cli/__snapshots__/model-fallback.test.ts.snap b/src/cli/__snapshots__/model-fallback.test.ts.snap index 285f14f5..86193b65 100644 --- a/src/cli/__snapshots__/model-fallback.test.ts.snap +++ b/src/cli/__snapshots__/model-fallback.test.ts.snap @@ -197,7 +197,7 @@ exports[`generateModelConfig single native provider uses OpenAI models when only "model": "opencode/gpt-5-nano", }, "hephaestus": { - "model": "openai/gpt-5.2-codex", + "model": "openai/gpt-5.3-codex", "variant": "medium", }, "librarian": { @@ -225,22 +225,22 @@ exports[`generateModelConfig single native provider uses OpenAI models when only }, "categories": { "deep": { - "model": "openai/gpt-5.2-codex", + "model": "openai/gpt-5.3-codex", "variant": "medium", }, "quick": { "model": "opencode/glm-4.7-free", }, "ultrabrain": { - "model": "openai/gpt-5.2-codex", + "model": "openai/gpt-5.3-codex", "variant": "xhigh", }, "unspecified-high": { - "model": "openai/gpt-5.2-codex", + "model": "openai/gpt-5.3-codex", "variant": "medium", }, "unspecified-low": { - "model": "openai/gpt-5.2-codex", + "model": "openai/gpt-5.3-codex", "variant": "medium", }, "visual-engineering": { @@ -264,7 +264,7 @@ exports[`generateModelConfig single native provider uses OpenAI models with isMa "model": "opencode/gpt-5-nano", }, "hephaestus": { - "model": "openai/gpt-5.2-codex", + "model": "openai/gpt-5.3-codex", "variant": "medium", }, "librarian": { @@ -292,14 +292,14 @@ exports[`generateModelConfig single native provider uses OpenAI models with isMa }, "categories": { "deep": { - "model": "openai/gpt-5.2-codex", + "model": "openai/gpt-5.3-codex", "variant": "medium", }, "quick": { "model": "opencode/glm-4.7-free", }, "ultrabrain": { - "model": "openai/gpt-5.2-codex", + "model": "openai/gpt-5.3-codex", "variant": "xhigh", }, "unspecified-high": { @@ -307,7 +307,7 @@ exports[`generateModelConfig single native provider uses OpenAI models with isMa "variant": "high", }, "unspecified-low": { - "model": "openai/gpt-5.2-codex", + "model": "openai/gpt-5.3-codex", "variant": "medium", }, "visual-engineering": { @@ -451,7 +451,7 @@ exports[`generateModelConfig all native providers uses preferred models from fal "model": "anthropic/claude-haiku-4-5", }, "hephaestus": { - "model": "openai/gpt-5.2-codex", + "model": "openai/gpt-5.3-codex", "variant": "medium", }, "librarian": { @@ -487,14 +487,14 @@ exports[`generateModelConfig all native providers uses preferred models from fal "variant": "high", }, "deep": { - "model": "openai/gpt-5.2-codex", + "model": "openai/gpt-5.3-codex", "variant": "medium", }, "quick": { "model": "anthropic/claude-haiku-4-5", }, "ultrabrain": { - "model": "openai/gpt-5.2-codex", + "model": "openai/gpt-5.3-codex", "variant": "xhigh", }, "unspecified-high": { @@ -524,7 +524,7 @@ exports[`generateModelConfig all native providers uses preferred models with isM "model": "anthropic/claude-haiku-4-5", }, "hephaestus": { - "model": "openai/gpt-5.2-codex", + "model": "openai/gpt-5.3-codex", "variant": "medium", }, "librarian": { @@ -560,14 +560,14 @@ exports[`generateModelConfig all native providers uses preferred models with isM "variant": "high", }, "deep": { - "model": "openai/gpt-5.2-codex", + "model": "openai/gpt-5.3-codex", "variant": "medium", }, "quick": { "model": "anthropic/claude-haiku-4-5", }, "ultrabrain": { - "model": "openai/gpt-5.2-codex", + "model": "openai/gpt-5.3-codex", "variant": "xhigh", }, "unspecified-high": { @@ -598,14 +598,14 @@ exports[`generateModelConfig fallback providers uses OpenCode Zen models when on "model": "opencode/claude-haiku-4-5", }, "hephaestus": { - "model": "opencode/gpt-5.2-codex", + "model": "opencode/gpt-5.3-codex", "variant": "medium", }, "librarian": { "model": "opencode/glm-4.7-free", }, "metis": { - "model": "opencode/claude-opus-4-5", + "model": "opencode/claude-opus-4-6", "variant": "max", }, "momus": { @@ -620,11 +620,11 @@ exports[`generateModelConfig fallback providers uses OpenCode Zen models when on "variant": "high", }, "prometheus": { - "model": "opencode/claude-opus-4-5", + "model": "opencode/claude-opus-4-6", "variant": "max", }, "sisyphus": { - "model": "opencode/claude-opus-4-5", + "model": "opencode/claude-opus-4-6", "variant": "max", }, }, @@ -634,14 +634,14 @@ exports[`generateModelConfig fallback providers uses OpenCode Zen models when on "variant": "high", }, "deep": { - "model": "opencode/gpt-5.2-codex", + "model": "opencode/gpt-5.3-codex", "variant": "medium", }, "quick": { "model": "opencode/claude-haiku-4-5", }, "ultrabrain": { - "model": "opencode/gpt-5.2-codex", + "model": "opencode/gpt-5.3-codex", "variant": "xhigh", }, "unspecified-high": { @@ -671,14 +671,14 @@ exports[`generateModelConfig fallback providers uses OpenCode Zen models with is "model": "opencode/claude-haiku-4-5", }, "hephaestus": { - "model": "opencode/gpt-5.2-codex", + "model": "opencode/gpt-5.3-codex", "variant": "medium", }, "librarian": { "model": "opencode/glm-4.7-free", }, "metis": { - "model": "opencode/claude-opus-4-5", + "model": "opencode/claude-opus-4-6", "variant": "max", }, "momus": { @@ -693,11 +693,11 @@ exports[`generateModelConfig fallback providers uses OpenCode Zen models with is "variant": "high", }, "prometheus": { - "model": "opencode/claude-opus-4-5", + "model": "opencode/claude-opus-4-6", "variant": "max", }, "sisyphus": { - "model": "opencode/claude-opus-4-5", + "model": "opencode/claude-opus-4-6", "variant": "max", }, }, @@ -707,18 +707,18 @@ exports[`generateModelConfig fallback providers uses OpenCode Zen models with is "variant": "high", }, "deep": { - "model": "opencode/gpt-5.2-codex", + "model": "opencode/gpt-5.3-codex", "variant": "medium", }, "quick": { "model": "opencode/claude-haiku-4-5", }, "ultrabrain": { - "model": "opencode/gpt-5.2-codex", + "model": "opencode/gpt-5.3-codex", "variant": "xhigh", }, "unspecified-high": { - "model": "opencode/claude-opus-4-5", + "model": "opencode/claude-opus-4-6", "variant": "max", }, "unspecified-low": { @@ -745,14 +745,14 @@ exports[`generateModelConfig fallback providers uses GitHub Copilot models when "model": "github-copilot/gpt-5-mini", }, "hephaestus": { - "model": "github-copilot/gpt-5.2-codex", + "model": "github-copilot/gpt-5.3-codex", "variant": "medium", }, "librarian": { "model": "github-copilot/claude-sonnet-4.5", }, "metis": { - "model": "github-copilot/claude-opus-4.5", + "model": "github-copilot/claude-opus-4.6", "variant": "max", }, "momus": { @@ -767,11 +767,11 @@ exports[`generateModelConfig fallback providers uses GitHub Copilot models when "variant": "high", }, "prometheus": { - "model": "github-copilot/claude-opus-4.5", + "model": "github-copilot/claude-opus-4.6", "variant": "max", }, "sisyphus": { - "model": "github-copilot/claude-opus-4.5", + "model": "github-copilot/claude-opus-4.6", "variant": "max", }, }, @@ -781,14 +781,14 @@ exports[`generateModelConfig fallback providers uses GitHub Copilot models when "variant": "high", }, "deep": { - "model": "github-copilot/gpt-5.2-codex", + "model": "github-copilot/gpt-5.3-codex", "variant": "medium", }, "quick": { "model": "github-copilot/claude-haiku-4.5", }, "ultrabrain": { - "model": "github-copilot/gpt-5.2-codex", + "model": "github-copilot/gpt-5.3-codex", "variant": "xhigh", }, "unspecified-high": { @@ -818,14 +818,14 @@ exports[`generateModelConfig fallback providers uses GitHub Copilot models with "model": "github-copilot/gpt-5-mini", }, "hephaestus": { - "model": "github-copilot/gpt-5.2-codex", + "model": "github-copilot/gpt-5.3-codex", "variant": "medium", }, "librarian": { "model": "github-copilot/claude-sonnet-4.5", }, "metis": { - "model": "github-copilot/claude-opus-4.5", + "model": "github-copilot/claude-opus-4.6", "variant": "max", }, "momus": { @@ -840,11 +840,11 @@ exports[`generateModelConfig fallback providers uses GitHub Copilot models with "variant": "high", }, "prometheus": { - "model": "github-copilot/claude-opus-4.5", + "model": "github-copilot/claude-opus-4.6", "variant": "max", }, "sisyphus": { - "model": "github-copilot/claude-opus-4.5", + "model": "github-copilot/claude-opus-4.6", "variant": "max", }, }, @@ -854,18 +854,18 @@ exports[`generateModelConfig fallback providers uses GitHub Copilot models with "variant": "high", }, "deep": { - "model": "github-copilot/gpt-5.2-codex", + "model": "github-copilot/gpt-5.3-codex", "variant": "medium", }, "quick": { "model": "github-copilot/claude-haiku-4.5", }, "ultrabrain": { - "model": "github-copilot/gpt-5.2-codex", + "model": "github-copilot/gpt-5.3-codex", "variant": "xhigh", }, "unspecified-high": { - "model": "github-copilot/claude-opus-4.5", + "model": "github-copilot/claude-opus-4.6", "variant": "max", }, "unspecified-low": { @@ -1002,7 +1002,7 @@ exports[`generateModelConfig mixed provider scenarios uses Claude + OpenCode Zen "model": "anthropic/claude-haiku-4-5", }, "hephaestus": { - "model": "opencode/gpt-5.2-codex", + "model": "opencode/gpt-5.3-codex", "variant": "medium", }, "librarian": { @@ -1038,14 +1038,14 @@ exports[`generateModelConfig mixed provider scenarios uses Claude + OpenCode Zen "variant": "high", }, "deep": { - "model": "opencode/gpt-5.2-codex", + "model": "opencode/gpt-5.3-codex", "variant": "medium", }, "quick": { "model": "anthropic/claude-haiku-4-5", }, "ultrabrain": { - "model": "opencode/gpt-5.2-codex", + "model": "opencode/gpt-5.3-codex", "variant": "xhigh", }, "unspecified-high": { @@ -1075,14 +1075,14 @@ exports[`generateModelConfig mixed provider scenarios uses OpenAI + Copilot comb "model": "github-copilot/gpt-5-mini", }, "hephaestus": { - "model": "openai/gpt-5.2-codex", + "model": "openai/gpt-5.3-codex", "variant": "medium", }, "librarian": { "model": "github-copilot/claude-sonnet-4.5", }, "metis": { - "model": "github-copilot/claude-opus-4.5", + "model": "github-copilot/claude-opus-4.6", "variant": "max", }, "momus": { @@ -1097,11 +1097,11 @@ exports[`generateModelConfig mixed provider scenarios uses OpenAI + Copilot comb "variant": "high", }, "prometheus": { - "model": "github-copilot/claude-opus-4.5", + "model": "github-copilot/claude-opus-4.6", "variant": "max", }, "sisyphus": { - "model": "github-copilot/claude-opus-4.5", + "model": "github-copilot/claude-opus-4.6", "variant": "max", }, }, @@ -1111,14 +1111,14 @@ exports[`generateModelConfig mixed provider scenarios uses OpenAI + Copilot comb "variant": "high", }, "deep": { - "model": "openai/gpt-5.2-codex", + "model": "openai/gpt-5.3-codex", "variant": "medium", }, "quick": { "model": "github-copilot/claude-haiku-4.5", }, "ultrabrain": { - "model": "openai/gpt-5.2-codex", + "model": "openai/gpt-5.3-codex", "variant": "xhigh", }, "unspecified-high": { @@ -1275,14 +1275,14 @@ exports[`generateModelConfig mixed provider scenarios uses all fallback provider "model": "opencode/claude-haiku-4-5", }, "hephaestus": { - "model": "github-copilot/gpt-5.2-codex", + "model": "github-copilot/gpt-5.3-codex", "variant": "medium", }, "librarian": { "model": "zai-coding-plan/glm-4.7", }, "metis": { - "model": "github-copilot/claude-opus-4.5", + "model": "github-copilot/claude-opus-4.6", "variant": "max", }, "momus": { @@ -1297,11 +1297,11 @@ exports[`generateModelConfig mixed provider scenarios uses all fallback provider "variant": "high", }, "prometheus": { - "model": "github-copilot/claude-opus-4.5", + "model": "github-copilot/claude-opus-4.6", "variant": "max", }, "sisyphus": { - "model": "github-copilot/claude-opus-4.5", + "model": "github-copilot/claude-opus-4.6", "variant": "max", }, }, @@ -1311,14 +1311,14 @@ exports[`generateModelConfig mixed provider scenarios uses all fallback provider "variant": "high", }, "deep": { - "model": "github-copilot/gpt-5.2-codex", + "model": "github-copilot/gpt-5.3-codex", "variant": "medium", }, "quick": { "model": "github-copilot/claude-haiku-4.5", }, "ultrabrain": { - "model": "github-copilot/gpt-5.2-codex", + "model": "github-copilot/gpt-5.3-codex", "variant": "xhigh", }, "unspecified-high": { @@ -1348,7 +1348,7 @@ exports[`generateModelConfig mixed provider scenarios uses all providers togethe "model": "anthropic/claude-haiku-4-5", }, "hephaestus": { - "model": "openai/gpt-5.2-codex", + "model": "openai/gpt-5.3-codex", "variant": "medium", }, "librarian": { @@ -1384,14 +1384,14 @@ exports[`generateModelConfig mixed provider scenarios uses all providers togethe "variant": "high", }, "deep": { - "model": "openai/gpt-5.2-codex", + "model": "openai/gpt-5.3-codex", "variant": "medium", }, "quick": { "model": "anthropic/claude-haiku-4-5", }, "ultrabrain": { - "model": "openai/gpt-5.2-codex", + "model": "openai/gpt-5.3-codex", "variant": "xhigh", }, "unspecified-high": { @@ -1421,7 +1421,7 @@ exports[`generateModelConfig mixed provider scenarios uses all providers with is "model": "anthropic/claude-haiku-4-5", }, "hephaestus": { - "model": "openai/gpt-5.2-codex", + "model": "openai/gpt-5.3-codex", "variant": "medium", }, "librarian": { @@ -1457,14 +1457,14 @@ exports[`generateModelConfig mixed provider scenarios uses all providers with is "variant": "high", }, "deep": { - "model": "openai/gpt-5.2-codex", + "model": "openai/gpt-5.3-codex", "variant": "medium", }, "quick": { "model": "anthropic/claude-haiku-4-5", }, "ultrabrain": { - "model": "openai/gpt-5.2-codex", + "model": "openai/gpt-5.3-codex", "variant": "xhigh", }, "unspecified-high": { diff --git a/src/hooks/think-mode/switcher.ts b/src/hooks/think-mode/switcher.ts index bec6f7d0..677c28b3 100644 --- a/src/hooks/think-mode/switcher.ts +++ b/src/hooks/think-mode/switcher.ts @@ -88,7 +88,6 @@ function resolveProvider(providerID: string, modelID: string): string { const HIGH_VARIANT_MAP: Record = { // Claude "claude-sonnet-4-5": "claude-sonnet-4-5-high", - "claude-opus-4-6": "claude-opus-4-6-high", "claude-opus-4-6": "claude-opus-4-6-high", // Gemini "gemini-3-pro": "gemini-3-pro-high",