From 1312d51366024825e731bd5976ba34924baeb7f4 Mon Sep 17 00:00:00 2001 From: Charm <124303983+charmcli@users.noreply.github.com> Date: Sat, 28 Mar 2026 03:06:37 +0000 Subject: [PATCH] chore: auto-update generated files --- internal/providers/configs/aihubmix.json | 19 ++ internal/providers/configs/ionet.json | 6 +- internal/providers/configs/openrouter.json | 276 ++++++++++----------- internal/providers/configs/synthetic.json | 22 +- internal/providers/configs/venice.json | 19 ++ internal/providers/configs/vercel.json | 19 ++ 6 files changed, 206 insertions(+), 155 deletions(-) diff --git a/internal/providers/configs/aihubmix.json b/internal/providers/configs/aihubmix.json index 9f99cf082314934068a93365243d67b3c971ebca..d3baf50e2832952caa1cd790c56321ee5f6ae434 100644 --- a/internal/providers/configs/aihubmix.json +++ b/internal/providers/configs/aihubmix.json @@ -226,6 +226,25 @@ "supports_attachments": true, "options": {} }, + { + "id": "anthropic-opus-4-6", + "name": "Anthropic Opus 4.6", + "cost_per_1m_in": 5, + "cost_per_1m_out": 25, + "cost_per_1m_in_cached": 6.25, + "cost_per_1m_out_cached": 0.5, + "context_window": 200000, + "default_max_tokens": 32000, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": true, + "options": {} + }, { "id": "claude-3-5-haiku", "name": "Claude 3.5 Haiku", diff --git a/internal/providers/configs/ionet.json b/internal/providers/configs/ionet.json index bae373d69e33dc89c949922efdebcfe470c70ea8..bddc72e3ba27e23fe5aedd49c12a7c42b40aa1f6 100644 --- a/internal/providers/configs/ionet.json +++ b/internal/providers/configs/ionet.json @@ -87,9 +87,9 @@ { "id": "moonshotai/Kimi-K2.5", "name": "MoonshotAI: Kimi K2.5", - "cost_per_1m_in": 0.55, - "cost_per_1m_out": 3, - "cost_per_1m_in_cached": 0.275, + "cost_per_1m_in": 0.45, + "cost_per_1m_out": 2.22, + "cost_per_1m_in_cached": 0.225, "cost_per_1m_out_cached": 1.1, "context_window": 262144, "default_max_tokens": 26214, diff --git a/internal/providers/configs/openrouter.json b/internal/providers/configs/openrouter.json index 09da2b6221c9f3df896ebe2f14e0d331b8412bf8..55407225318a4d2356c2939d5af5de8082ed1ff3 100644 --- a/internal/providers/configs/openrouter.json +++ b/internal/providers/configs/openrouter.json @@ -560,10 +560,10 @@ { "id": "deepseek/deepseek-chat-v3.1", "name": "DeepSeek: DeepSeek V3.1", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 1.7, + "cost_per_1m_in": 0.56, + "cost_per_1m_out": 1.68, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.28, "context_window": 163840, "default_max_tokens": 16384, "can_reason": true, @@ -598,10 +598,10 @@ { "id": "deepseek/deepseek-v3.2", "name": "DeepSeek: DeepSeek V3.2", - "cost_per_1m_in": 0.5, - "cost_per_1m_out": 1.5, + "cost_per_1m_in": 0.26, + "cost_per_1m_out": 0.38, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.25, + "cost_per_1m_out_cached": 0.13, "context_window": 163840, "default_max_tokens": 81920, "can_reason": true, @@ -655,10 +655,10 @@ { "id": "deepseek/deepseek-r1-0528", "name": "DeepSeek: R1 0528", - "cost_per_1m_in": 0.8, - "cost_per_1m_out": 2.4, + "cost_per_1m_in": 0.7, + "cost_per_1m_out": 2.5, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.35, "context_window": 163840, "default_max_tokens": 16384, "can_reason": true, @@ -832,7 +832,7 @@ "cost_per_1m_in_cached": 0.08333, "cost_per_1m_out_cached": 0.05, "context_window": 1048576, - "default_max_tokens": 32768, + "default_max_tokens": 32767, "can_reason": true, "reasoning_levels": [ "low", @@ -958,6 +958,19 @@ "supports_attachments": false, "options": {} }, + { + "id": "kwaipilot/kat-coder-pro-v2", + "name": "Kwaipilot: KAT-Coder-Pro V2", + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 1.2, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0.06, + "context_window": 256000, + "default_max_tokens": 40000, + "can_reason": false, + "supports_attachments": false, + "options": {} + }, { "id": "meituan/longcat-flash-chat", "name": "Meituan: LongCat Flash Chat", @@ -979,7 +992,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 8192, + "default_max_tokens": 13107, "can_reason": false, "supports_attachments": false, "options": {} @@ -987,12 +1000,12 @@ { "id": "meta-llama/llama-3.3-70b-instruct", "name": "Meta: Llama 3.3 70B Instruct", - "cost_per_1m_in": 0.13, - "cost_per_1m_out": 0.4, + "cost_per_1m_in": 0.25, + "cost_per_1m_out": 0.75, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 64000, + "default_max_tokens": 13107, "can_reason": false, "supports_attachments": false, "options": {} @@ -1036,34 +1049,15 @@ "supports_attachments": true, "options": {} }, - { - "id": "minimax/minimax-m1", - "name": "MiniMax: MiniMax M1", - "cost_per_1m_in": 0.44, - "cost_per_1m_out": 1.76, - "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 1000000, - "default_max_tokens": 20000, - "can_reason": true, - "reasoning_levels": [ - "low", - "medium", - "high" - ], - "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} - }, { "id": "minimax/minimax-m2", "name": "MiniMax: MiniMax M2", "cost_per_1m_in": 0.255, - "cost_per_1m_out": 1.02, + "cost_per_1m_out": 1, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 204800, - "default_max_tokens": 65536, + "cost_per_1m_out_cached": 0.03, + "context_window": 196608, + "default_max_tokens": 98304, "can_reason": true, "reasoning_levels": [ "low", @@ -1077,10 +1071,10 @@ { "id": "minimax/minimax-m2.1", "name": "MiniMax: MiniMax M2.1", - "cost_per_1m_in": 0.3, - "cost_per_1m_out": 1.2, + "cost_per_1m_in": 0.27, + "cost_per_1m_out": 0.95, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.029, "context_window": 196608, "default_max_tokens": 19660, "can_reason": true, @@ -1134,8 +1128,8 @@ { "id": "minimax/minimax-m2.7", "name": "MiniMax: MiniMax M2.7", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 2.4, + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 1.2, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.06, "context_window": 204800, @@ -1244,12 +1238,12 @@ { "id": "mistralai/ministral-14b-2512", "name": "Mistral: Ministral 3 14B 2512", - "cost_per_1m_in": 0.35, - "cost_per_1m_out": 0.35, + "cost_per_1m_in": 0.2, + "cost_per_1m_out": 0.2, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.02, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 26214, "can_reason": false, "supports_attachments": true, "options": {} @@ -1257,12 +1251,12 @@ { "id": "mistralai/ministral-3b-2512", "name": "Mistral: Ministral 3 3B 2512", - "cost_per_1m_in": 0.35, - "cost_per_1m_out": 0.35, + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.1, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.01, "context_window": 131072, - "default_max_tokens": 65536, + "default_max_tokens": 13107, "can_reason": false, "supports_attachments": true, "options": {} @@ -1270,12 +1264,12 @@ { "id": "mistralai/ministral-8b-2512", "name": "Mistral: Ministral 3 8B 2512", - "cost_per_1m_in": 0.3, - "cost_per_1m_out": 0.3, + "cost_per_1m_in": 0.15, + "cost_per_1m_out": 0.15, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.015, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 26214, "can_reason": false, "supports_attachments": true, "options": {} @@ -1487,9 +1481,9 @@ "cost_per_1m_in": 0.6, "cost_per_1m_out": 2.5, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.6, + "cost_per_1m_out_cached": 0.15, "context_window": 262144, - "default_max_tokens": 16384, + "default_max_tokens": 26214, "can_reason": false, "supports_attachments": false, "options": {} @@ -1497,12 +1491,12 @@ { "id": "moonshotai/kimi-k2-thinking", "name": "MoonshotAI: Kimi K2 Thinking", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 2.5, + "cost_per_1m_in": 0.47, + "cost_per_1m_out": 2, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 262144, - "default_max_tokens": 26214, + "cost_per_1m_out_cached": 0.141, + "context_window": 131072, + "default_max_tokens": 13107, "can_reason": true, "reasoning_levels": [ "low", @@ -1516,12 +1510,12 @@ { "id": "moonshotai/kimi-k2.5", "name": "MoonshotAI: Kimi K2.5", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 3, + "cost_per_1m_in": 0.45, + "cost_per_1m_out": 2.25, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.1, + "cost_per_1m_out_cached": 0.07, "context_window": 262144, - "default_max_tokens": 26214, + "default_max_tokens": 131072, "can_reason": true, "reasoning_levels": [ "low", @@ -1602,6 +1596,25 @@ "supports_attachments": false, "options": {} }, + { + "id": "nvidia/nemotron-3-super-120b-a12b", + "name": "NVIDIA: Nemotron 3 Super", + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.5, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0.1, + "context_window": 262144, + "default_max_tokens": 26214, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false, + "options": {} + }, { "id": "nvidia/nemotron-3-super-120b-a12b:free", "name": "NVIDIA: Nemotron 3 Super (free)", @@ -2031,9 +2044,9 @@ "cost_per_1m_in": 1.25, "cost_per_1m_out": 10, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.13, + "cost_per_1m_out_cached": 0.125, "context_window": 128000, - "default_max_tokens": 16000, + "default_max_tokens": 8192, "can_reason": false, "supports_attachments": true, "options": {} @@ -2276,12 +2289,12 @@ { "id": "openai/gpt-oss-120b", "name": "OpenAI: gpt-oss-120b", - "cost_per_1m_in": 0.1, - "cost_per_1m_out": 0.49, + "cost_per_1m_in": 0.05, + "cost_per_1m_out": 0.25, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 65536, + "default_max_tokens": 16384, "can_reason": true, "reasoning_levels": [ "low", @@ -2687,25 +2700,6 @@ "supports_attachments": false, "options": {} }, - { - "id": "qwen/qwen3-14b", - "name": "Qwen: Qwen3 14B", - "cost_per_1m_in": 0.2275, - "cost_per_1m_out": 0.91, - "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 131072, - "default_max_tokens": 4096, - "can_reason": true, - "reasoning_levels": [ - "low", - "medium", - "high" - ], - "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} - }, { "id": "qwen/qwen3-235b-a22b", "name": "Qwen: Qwen3 235B A22B", @@ -2760,12 +2754,12 @@ { "id": "qwen/qwen3-30b-a3b", "name": "Qwen: Qwen3 30B A3B", - "cost_per_1m_in": 0.15, - "cost_per_1m_out": 0.6, + "cost_per_1m_in": 0.13, + "cost_per_1m_out": 0.52, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.15, + "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 65536, + "default_max_tokens": 4096, "can_reason": true, "reasoning_levels": [ "low", @@ -2792,12 +2786,12 @@ { "id": "qwen/qwen3-30b-a3b-thinking-2507", "name": "Qwen: Qwen3 30B A3B Thinking 2507", - "cost_per_1m_in": 0.1, - "cost_per_1m_out": 0.3, + "cost_per_1m_in": 0.08, + "cost_per_1m_out": 0.4, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 262144, - "default_max_tokens": 26214, + "cost_per_1m_out_cached": 0.08, + "context_window": 131072, + "default_max_tokens": 65536, "can_reason": true, "reasoning_levels": [ "low", @@ -2920,12 +2914,12 @@ { "id": "qwen/qwen3-coder-next", "name": "Qwen: Qwen3 Coder Next", - "cost_per_1m_in": 0.18, - "cost_per_1m_out": 1.35, + "cost_per_1m_in": 0.12, + "cost_per_1m_out": 0.75, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.18, + "cost_per_1m_out_cached": 0.06, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 32768, "can_reason": false, "supports_attachments": false, "options": {} @@ -3023,12 +3017,12 @@ { "id": "qwen/qwen3-vl-235b-a22b-instruct", "name": "Qwen: Qwen3 VL 235B A22B Instruct", - "cost_per_1m_in": 0.26, - "cost_per_1m_out": 1.04, + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 1.5, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 131072, - "default_max_tokens": 16384, + "context_window": 262144, + "default_max_tokens": 131072, "can_reason": false, "supports_attachments": true, "options": {} @@ -3055,12 +3049,12 @@ { "id": "qwen/qwen3-vl-30b-a3b-instruct", "name": "Qwen: Qwen3 VL 30B A3B Instruct", - "cost_per_1m_in": 0.13, - "cost_per_1m_out": 0.52, + "cost_per_1m_in": 0.29, + "cost_per_1m_out": 1, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 131072, - "default_max_tokens": 16384, + "context_window": 262144, + "default_max_tokens": 131072, "can_reason": false, "supports_attachments": true, "options": {} @@ -3132,12 +3126,12 @@ { "id": "qwen/qwen3.5-397b-a17b", "name": "Qwen: Qwen3.5 397B A17B", - "cost_per_1m_in": 0.55, - "cost_per_1m_out": 3.5, + "cost_per_1m_in": 0.6, + "cost_per_1m_out": 3.6, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.55, + "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 32768, + "default_max_tokens": 131072, "can_reason": true, "reasoning_levels": [ "low", @@ -3170,10 +3164,10 @@ { "id": "qwen/qwen3.5-122b-a10b", "name": "Qwen: Qwen3.5-122B-A10B", - "cost_per_1m_in": 0.3, - "cost_per_1m_out": 2.4, + "cost_per_1m_in": 0.26, + "cost_per_1m_out": 2.08, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.3, + "cost_per_1m_out_cached": 0, "context_window": 262144, "default_max_tokens": 32768, "can_reason": true, @@ -3189,10 +3183,10 @@ { "id": "qwen/qwen3.5-27b", "name": "Qwen: Qwen3.5-27B", - "cost_per_1m_in": 0.3, - "cost_per_1m_out": 2.4, + "cost_per_1m_in": 0.27, + "cost_per_1m_out": 2.16, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.27, "context_window": 262144, "default_max_tokens": 32768, "can_reason": true, @@ -3488,12 +3482,12 @@ { "id": "z-ai/glm-4.5-air", "name": "Z.ai: GLM 4.5 Air", - "cost_per_1m_in": 0.14, - "cost_per_1m_out": 0.86, + "cost_per_1m_in": 0.2, + "cost_per_1m_out": 1.2, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 65536, + "default_max_tokens": 13107, "can_reason": true, "reasoning_levels": [ "low", @@ -3545,12 +3539,12 @@ { "id": "z-ai/glm-4.6", "name": "Z.ai: GLM 4.6", - "cost_per_1m_in": 0.55, - "cost_per_1m_out": 2.2, + "cost_per_1m_in": 0.39, + "cost_per_1m_out": 1.9, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.11, + "cost_per_1m_out_cached": 0, "context_window": 204800, - "default_max_tokens": 65536, + "default_max_tokens": 102400, "can_reason": true, "reasoning_levels": [ "low", @@ -3567,9 +3561,9 @@ "cost_per_1m_in": 0.3, "cost_per_1m_out": 0.9, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.05, "context_window": 131072, - "default_max_tokens": 13107, + "default_max_tokens": 12000, "can_reason": true, "reasoning_levels": [ "low", @@ -3659,10 +3653,10 @@ { "id": "x-ai/grok-3", "name": "xAI: Grok 3", - "cost_per_1m_in": 3, - "cost_per_1m_out": 15, + "cost_per_1m_in": 5, + "cost_per_1m_out": 25, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.75, + "cost_per_1m_out_cached": 1.25, "context_window": 131072, "default_max_tokens": 13107, "can_reason": false, @@ -3672,10 +3666,10 @@ { "id": "x-ai/grok-3-beta", "name": "xAI: Grok 3 Beta", - "cost_per_1m_in": 3, - "cost_per_1m_out": 15, + "cost_per_1m_in": 5, + "cost_per_1m_out": 25, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.75, + "cost_per_1m_out_cached": 1.25, "context_window": 131072, "default_max_tokens": 13107, "can_reason": false, @@ -3685,10 +3679,10 @@ { "id": "x-ai/grok-3-mini", "name": "xAI: Grok 3 Mini", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 4, + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 0.5, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.15, + "cost_per_1m_out_cached": 0.075, "context_window": 131072, "default_max_tokens": 13107, "can_reason": true, @@ -3704,10 +3698,10 @@ { "id": "x-ai/grok-3-mini-beta", "name": "xAI: Grok 3 Mini Beta", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 4, + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 0.5, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.15, + "cost_per_1m_out_cached": 0.075, "context_window": 131072, "default_max_tokens": 13107, "can_reason": true, diff --git a/internal/providers/configs/synthetic.json b/internal/providers/configs/synthetic.json index 0ef0c5759f86acdc35e0579e893951e555c13b71..7c4eacea815dd236dcbe2487c6d2afed9f36a2a1 100644 --- a/internal/providers/configs/synthetic.json +++ b/internal/providers/configs/synthetic.json @@ -61,10 +61,10 @@ { "id": "hf:zai-org/GLM-4.7", "name": "GLM 4.7", - "cost_per_1m_in": 0.55, + "cost_per_1m_in": 0.45, "cost_per_1m_out": 2.19, - "cost_per_1m_in_cached": 0.55, - "cost_per_1m_out_cached": 0.55, + "cost_per_1m_in_cached": 0.45, + "cost_per_1m_out_cached": 0.45, "context_window": 202752, "default_max_tokens": 20275, "can_reason": true, @@ -131,10 +131,10 @@ { "id": "hf:moonshotai/Kimi-K2.5", "name": "Kimi K2.5", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 3, - "cost_per_1m_in_cached": 0.6, - "cost_per_1m_out_cached": 0.6, + "cost_per_1m_in": 0.45, + "cost_per_1m_out": 3.4, + "cost_per_1m_in_cached": 0.45, + "cost_per_1m_out_cached": 0.45, "context_window": 262144, "default_max_tokens": 32768, "can_reason": true, @@ -150,10 +150,10 @@ { "id": "hf:nvidia/Kimi-K2.5-NVFP4", "name": "Kimi K2.5 NVFP4", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 3, - "cost_per_1m_in_cached": 0.6, - "cost_per_1m_out_cached": 0.6, + "cost_per_1m_in": 0.45, + "cost_per_1m_out": 3.4, + "cost_per_1m_in_cached": 0.45, + "cost_per_1m_out_cached": 0.45, "context_window": 262144, "default_max_tokens": 32768, "can_reason": true, diff --git a/internal/providers/configs/venice.json b/internal/providers/configs/venice.json index bcab929df8b5be0404b58184edde3e3bceed0479..830515f23ea3dfe14436346a67133ceab20b6cf8 100644 --- a/internal/providers/configs/venice.json +++ b/internal/providers/configs/venice.json @@ -293,6 +293,25 @@ "supports_attachments": true, "options": {} }, + { + "id": "openai-gpt-54-mini", + "name": "GPT-5.4 Mini", + "cost_per_1m_in": 0.9375, + "cost_per_1m_out": 5.625, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 400000, + "default_max_tokens": 32768, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": true, + "options": {} + }, { "id": "openai-gpt-54-pro", "name": "GPT-5.4 Pro", diff --git a/internal/providers/configs/vercel.json b/internal/providers/configs/vercel.json index 7f82a70e21acaee5e444317a068bdfa569fe1dc2..6472fa7d338df32079255ea9efbad80efdc4275b 100644 --- a/internal/providers/configs/vercel.json +++ b/internal/providers/configs/vercel.json @@ -1538,6 +1538,25 @@ "supports_attachments": false, "options": {} }, + { + "id": "kwaipilot/kat-coder-pro-v2", + "name": "Kat Coder Pro V2", + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 1.2, + "cost_per_1m_in_cached": 0.06, + "cost_per_1m_out_cached": 0, + "context_window": 256000, + "default_max_tokens": 8000, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false, + "options": {} + }, { "id": "moonshotai/kimi-k2", "name": "Kimi K2",