fix: update snapshots and remove duplicate key in switcher for model version update

This commit is contained in:
YeonGyu-Kim
2026-02-06 15:12:41 +09:00
parent 1f64920453
commit 25e436a4aa
2 changed files with 62 additions and 63 deletions

View File

@@ -197,7 +197,7 @@ exports[`generateModelConfig single native provider uses OpenAI models when only
"model": "opencode/gpt-5-nano",
},
"hephaestus": {
"model": "openai/gpt-5.2-codex",
"model": "openai/gpt-5.3-codex",
"variant": "medium",
},
"librarian": {
@@ -225,22 +225,22 @@ exports[`generateModelConfig single native provider uses OpenAI models when only
},
"categories": {
"deep": {
"model": "openai/gpt-5.2-codex",
"model": "openai/gpt-5.3-codex",
"variant": "medium",
},
"quick": {
"model": "opencode/glm-4.7-free",
},
"ultrabrain": {
"model": "openai/gpt-5.2-codex",
"model": "openai/gpt-5.3-codex",
"variant": "xhigh",
},
"unspecified-high": {
"model": "openai/gpt-5.2-codex",
"model": "openai/gpt-5.3-codex",
"variant": "medium",
},
"unspecified-low": {
"model": "openai/gpt-5.2-codex",
"model": "openai/gpt-5.3-codex",
"variant": "medium",
},
"visual-engineering": {
@@ -264,7 +264,7 @@ exports[`generateModelConfig single native provider uses OpenAI models with isMa
"model": "opencode/gpt-5-nano",
},
"hephaestus": {
"model": "openai/gpt-5.2-codex",
"model": "openai/gpt-5.3-codex",
"variant": "medium",
},
"librarian": {
@@ -292,14 +292,14 @@ exports[`generateModelConfig single native provider uses OpenAI models with isMa
},
"categories": {
"deep": {
"model": "openai/gpt-5.2-codex",
"model": "openai/gpt-5.3-codex",
"variant": "medium",
},
"quick": {
"model": "opencode/glm-4.7-free",
},
"ultrabrain": {
"model": "openai/gpt-5.2-codex",
"model": "openai/gpt-5.3-codex",
"variant": "xhigh",
},
"unspecified-high": {
@@ -307,7 +307,7 @@ exports[`generateModelConfig single native provider uses OpenAI models with isMa
"variant": "high",
},
"unspecified-low": {
"model": "openai/gpt-5.2-codex",
"model": "openai/gpt-5.3-codex",
"variant": "medium",
},
"visual-engineering": {
@@ -451,7 +451,7 @@ exports[`generateModelConfig all native providers uses preferred models from fal
"model": "anthropic/claude-haiku-4-5",
},
"hephaestus": {
"model": "openai/gpt-5.2-codex",
"model": "openai/gpt-5.3-codex",
"variant": "medium",
},
"librarian": {
@@ -487,14 +487,14 @@ exports[`generateModelConfig all native providers uses preferred models from fal
"variant": "high",
},
"deep": {
"model": "openai/gpt-5.2-codex",
"model": "openai/gpt-5.3-codex",
"variant": "medium",
},
"quick": {
"model": "anthropic/claude-haiku-4-5",
},
"ultrabrain": {
"model": "openai/gpt-5.2-codex",
"model": "openai/gpt-5.3-codex",
"variant": "xhigh",
},
"unspecified-high": {
@@ -524,7 +524,7 @@ exports[`generateModelConfig all native providers uses preferred models with isM
"model": "anthropic/claude-haiku-4-5",
},
"hephaestus": {
"model": "openai/gpt-5.2-codex",
"model": "openai/gpt-5.3-codex",
"variant": "medium",
},
"librarian": {
@@ -560,14 +560,14 @@ exports[`generateModelConfig all native providers uses preferred models with isM
"variant": "high",
},
"deep": {
"model": "openai/gpt-5.2-codex",
"model": "openai/gpt-5.3-codex",
"variant": "medium",
},
"quick": {
"model": "anthropic/claude-haiku-4-5",
},
"ultrabrain": {
"model": "openai/gpt-5.2-codex",
"model": "openai/gpt-5.3-codex",
"variant": "xhigh",
},
"unspecified-high": {
@@ -598,14 +598,14 @@ exports[`generateModelConfig fallback providers uses OpenCode Zen models when on
"model": "opencode/claude-haiku-4-5",
},
"hephaestus": {
"model": "opencode/gpt-5.2-codex",
"model": "opencode/gpt-5.3-codex",
"variant": "medium",
},
"librarian": {
"model": "opencode/glm-4.7-free",
},
"metis": {
"model": "opencode/claude-opus-4-5",
"model": "opencode/claude-opus-4-6",
"variant": "max",
},
"momus": {
@@ -620,11 +620,11 @@ exports[`generateModelConfig fallback providers uses OpenCode Zen models when on
"variant": "high",
},
"prometheus": {
"model": "opencode/claude-opus-4-5",
"model": "opencode/claude-opus-4-6",
"variant": "max",
},
"sisyphus": {
"model": "opencode/claude-opus-4-5",
"model": "opencode/claude-opus-4-6",
"variant": "max",
},
},
@@ -634,14 +634,14 @@ exports[`generateModelConfig fallback providers uses OpenCode Zen models when on
"variant": "high",
},
"deep": {
"model": "opencode/gpt-5.2-codex",
"model": "opencode/gpt-5.3-codex",
"variant": "medium",
},
"quick": {
"model": "opencode/claude-haiku-4-5",
},
"ultrabrain": {
"model": "opencode/gpt-5.2-codex",
"model": "opencode/gpt-5.3-codex",
"variant": "xhigh",
},
"unspecified-high": {
@@ -671,14 +671,14 @@ exports[`generateModelConfig fallback providers uses OpenCode Zen models with is
"model": "opencode/claude-haiku-4-5",
},
"hephaestus": {
"model": "opencode/gpt-5.2-codex",
"model": "opencode/gpt-5.3-codex",
"variant": "medium",
},
"librarian": {
"model": "opencode/glm-4.7-free",
},
"metis": {
"model": "opencode/claude-opus-4-5",
"model": "opencode/claude-opus-4-6",
"variant": "max",
},
"momus": {
@@ -693,11 +693,11 @@ exports[`generateModelConfig fallback providers uses OpenCode Zen models with is
"variant": "high",
},
"prometheus": {
"model": "opencode/claude-opus-4-5",
"model": "opencode/claude-opus-4-6",
"variant": "max",
},
"sisyphus": {
"model": "opencode/claude-opus-4-5",
"model": "opencode/claude-opus-4-6",
"variant": "max",
},
},
@@ -707,18 +707,18 @@ exports[`generateModelConfig fallback providers uses OpenCode Zen models with is
"variant": "high",
},
"deep": {
"model": "opencode/gpt-5.2-codex",
"model": "opencode/gpt-5.3-codex",
"variant": "medium",
},
"quick": {
"model": "opencode/claude-haiku-4-5",
},
"ultrabrain": {
"model": "opencode/gpt-5.2-codex",
"model": "opencode/gpt-5.3-codex",
"variant": "xhigh",
},
"unspecified-high": {
"model": "opencode/claude-opus-4-5",
"model": "opencode/claude-opus-4-6",
"variant": "max",
},
"unspecified-low": {
@@ -745,14 +745,14 @@ exports[`generateModelConfig fallback providers uses GitHub Copilot models when
"model": "github-copilot/gpt-5-mini",
},
"hephaestus": {
"model": "github-copilot/gpt-5.2-codex",
"model": "github-copilot/gpt-5.3-codex",
"variant": "medium",
},
"librarian": {
"model": "github-copilot/claude-sonnet-4.5",
},
"metis": {
"model": "github-copilot/claude-opus-4.5",
"model": "github-copilot/claude-opus-4.6",
"variant": "max",
},
"momus": {
@@ -767,11 +767,11 @@ exports[`generateModelConfig fallback providers uses GitHub Copilot models when
"variant": "high",
},
"prometheus": {
"model": "github-copilot/claude-opus-4.5",
"model": "github-copilot/claude-opus-4.6",
"variant": "max",
},
"sisyphus": {
"model": "github-copilot/claude-opus-4.5",
"model": "github-copilot/claude-opus-4.6",
"variant": "max",
},
},
@@ -781,14 +781,14 @@ exports[`generateModelConfig fallback providers uses GitHub Copilot models when
"variant": "high",
},
"deep": {
"model": "github-copilot/gpt-5.2-codex",
"model": "github-copilot/gpt-5.3-codex",
"variant": "medium",
},
"quick": {
"model": "github-copilot/claude-haiku-4.5",
},
"ultrabrain": {
"model": "github-copilot/gpt-5.2-codex",
"model": "github-copilot/gpt-5.3-codex",
"variant": "xhigh",
},
"unspecified-high": {
@@ -818,14 +818,14 @@ exports[`generateModelConfig fallback providers uses GitHub Copilot models with
"model": "github-copilot/gpt-5-mini",
},
"hephaestus": {
"model": "github-copilot/gpt-5.2-codex",
"model": "github-copilot/gpt-5.3-codex",
"variant": "medium",
},
"librarian": {
"model": "github-copilot/claude-sonnet-4.5",
},
"metis": {
"model": "github-copilot/claude-opus-4.5",
"model": "github-copilot/claude-opus-4.6",
"variant": "max",
},
"momus": {
@@ -840,11 +840,11 @@ exports[`generateModelConfig fallback providers uses GitHub Copilot models with
"variant": "high",
},
"prometheus": {
"model": "github-copilot/claude-opus-4.5",
"model": "github-copilot/claude-opus-4.6",
"variant": "max",
},
"sisyphus": {
"model": "github-copilot/claude-opus-4.5",
"model": "github-copilot/claude-opus-4.6",
"variant": "max",
},
},
@@ -854,18 +854,18 @@ exports[`generateModelConfig fallback providers uses GitHub Copilot models with
"variant": "high",
},
"deep": {
"model": "github-copilot/gpt-5.2-codex",
"model": "github-copilot/gpt-5.3-codex",
"variant": "medium",
},
"quick": {
"model": "github-copilot/claude-haiku-4.5",
},
"ultrabrain": {
"model": "github-copilot/gpt-5.2-codex",
"model": "github-copilot/gpt-5.3-codex",
"variant": "xhigh",
},
"unspecified-high": {
"model": "github-copilot/claude-opus-4.5",
"model": "github-copilot/claude-opus-4.6",
"variant": "max",
},
"unspecified-low": {
@@ -1002,7 +1002,7 @@ exports[`generateModelConfig mixed provider scenarios uses Claude + OpenCode Zen
"model": "anthropic/claude-haiku-4-5",
},
"hephaestus": {
"model": "opencode/gpt-5.2-codex",
"model": "opencode/gpt-5.3-codex",
"variant": "medium",
},
"librarian": {
@@ -1038,14 +1038,14 @@ exports[`generateModelConfig mixed provider scenarios uses Claude + OpenCode Zen
"variant": "high",
},
"deep": {
"model": "opencode/gpt-5.2-codex",
"model": "opencode/gpt-5.3-codex",
"variant": "medium",
},
"quick": {
"model": "anthropic/claude-haiku-4-5",
},
"ultrabrain": {
"model": "opencode/gpt-5.2-codex",
"model": "opencode/gpt-5.3-codex",
"variant": "xhigh",
},
"unspecified-high": {
@@ -1075,14 +1075,14 @@ exports[`generateModelConfig mixed provider scenarios uses OpenAI + Copilot comb
"model": "github-copilot/gpt-5-mini",
},
"hephaestus": {
"model": "openai/gpt-5.2-codex",
"model": "openai/gpt-5.3-codex",
"variant": "medium",
},
"librarian": {
"model": "github-copilot/claude-sonnet-4.5",
},
"metis": {
"model": "github-copilot/claude-opus-4.5",
"model": "github-copilot/claude-opus-4.6",
"variant": "max",
},
"momus": {
@@ -1097,11 +1097,11 @@ exports[`generateModelConfig mixed provider scenarios uses OpenAI + Copilot comb
"variant": "high",
},
"prometheus": {
"model": "github-copilot/claude-opus-4.5",
"model": "github-copilot/claude-opus-4.6",
"variant": "max",
},
"sisyphus": {
"model": "github-copilot/claude-opus-4.5",
"model": "github-copilot/claude-opus-4.6",
"variant": "max",
},
},
@@ -1111,14 +1111,14 @@ exports[`generateModelConfig mixed provider scenarios uses OpenAI + Copilot comb
"variant": "high",
},
"deep": {
"model": "openai/gpt-5.2-codex",
"model": "openai/gpt-5.3-codex",
"variant": "medium",
},
"quick": {
"model": "github-copilot/claude-haiku-4.5",
},
"ultrabrain": {
"model": "openai/gpt-5.2-codex",
"model": "openai/gpt-5.3-codex",
"variant": "xhigh",
},
"unspecified-high": {
@@ -1275,14 +1275,14 @@ exports[`generateModelConfig mixed provider scenarios uses all fallback provider
"model": "opencode/claude-haiku-4-5",
},
"hephaestus": {
"model": "github-copilot/gpt-5.2-codex",
"model": "github-copilot/gpt-5.3-codex",
"variant": "medium",
},
"librarian": {
"model": "zai-coding-plan/glm-4.7",
},
"metis": {
"model": "github-copilot/claude-opus-4.5",
"model": "github-copilot/claude-opus-4.6",
"variant": "max",
},
"momus": {
@@ -1297,11 +1297,11 @@ exports[`generateModelConfig mixed provider scenarios uses all fallback provider
"variant": "high",
},
"prometheus": {
"model": "github-copilot/claude-opus-4.5",
"model": "github-copilot/claude-opus-4.6",
"variant": "max",
},
"sisyphus": {
"model": "github-copilot/claude-opus-4.5",
"model": "github-copilot/claude-opus-4.6",
"variant": "max",
},
},
@@ -1311,14 +1311,14 @@ exports[`generateModelConfig mixed provider scenarios uses all fallback provider
"variant": "high",
},
"deep": {
"model": "github-copilot/gpt-5.2-codex",
"model": "github-copilot/gpt-5.3-codex",
"variant": "medium",
},
"quick": {
"model": "github-copilot/claude-haiku-4.5",
},
"ultrabrain": {
"model": "github-copilot/gpt-5.2-codex",
"model": "github-copilot/gpt-5.3-codex",
"variant": "xhigh",
},
"unspecified-high": {
@@ -1348,7 +1348,7 @@ exports[`generateModelConfig mixed provider scenarios uses all providers togethe
"model": "anthropic/claude-haiku-4-5",
},
"hephaestus": {
"model": "openai/gpt-5.2-codex",
"model": "openai/gpt-5.3-codex",
"variant": "medium",
},
"librarian": {
@@ -1384,14 +1384,14 @@ exports[`generateModelConfig mixed provider scenarios uses all providers togethe
"variant": "high",
},
"deep": {
"model": "openai/gpt-5.2-codex",
"model": "openai/gpt-5.3-codex",
"variant": "medium",
},
"quick": {
"model": "anthropic/claude-haiku-4-5",
},
"ultrabrain": {
"model": "openai/gpt-5.2-codex",
"model": "openai/gpt-5.3-codex",
"variant": "xhigh",
},
"unspecified-high": {
@@ -1421,7 +1421,7 @@ exports[`generateModelConfig mixed provider scenarios uses all providers with is
"model": "anthropic/claude-haiku-4-5",
},
"hephaestus": {
"model": "openai/gpt-5.2-codex",
"model": "openai/gpt-5.3-codex",
"variant": "medium",
},
"librarian": {
@@ -1457,14 +1457,14 @@ exports[`generateModelConfig mixed provider scenarios uses all providers with is
"variant": "high",
},
"deep": {
"model": "openai/gpt-5.2-codex",
"model": "openai/gpt-5.3-codex",
"variant": "medium",
},
"quick": {
"model": "anthropic/claude-haiku-4-5",
},
"ultrabrain": {
"model": "openai/gpt-5.2-codex",
"model": "openai/gpt-5.3-codex",
"variant": "xhigh",
},
"unspecified-high": {

View File

@@ -88,7 +88,6 @@ function resolveProvider(providerID: string, modelID: string): string {
const HIGH_VARIANT_MAP: Record<string, string> = {
// Claude
"claude-sonnet-4-5": "claude-sonnet-4-5-high",
"claude-opus-4-6": "claude-opus-4-6-high",
"claude-opus-4-6": "claude-opus-4-6-high",
// Gemini
"gemini-3-pro": "gemini-3-pro-high",