From 5a0bdb3322d383f1be70982f100e2d3e8906fcc0 Mon Sep 17 00:00:00 2001 From: Charm <124303983+charmcli@users.noreply.github.com> Date: Mon, 16 Mar 2026 18:09:45 +0000 Subject: [PATCH] chore: auto-update generated files --- internal/providers/configs/aihubmix.json | 76 ++++++++ internal/providers/configs/openrouter.json | 191 +++++++++++---------- internal/providers/configs/vercel.json | 19 ++ 3 files changed, 200 insertions(+), 86 deletions(-) diff --git a/internal/providers/configs/aihubmix.json b/internal/providers/configs/aihubmix.json index b90f951fb38ab4525b2c3496a078fc0425e70a09..50aec0fd374fe5231b81e18218e7dc1296f3c081 100644 --- a/internal/providers/configs/aihubmix.json +++ b/internal/providers/configs/aihubmix.json @@ -1702,6 +1702,25 @@ "supports_attachments": true, "options": {} }, + { + "id": "gpt-5-pro", + "name": "GPT 5 Pro", + "cost_per_1m_in": 15, + "cost_per_1m_out": 120, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 400000, + "default_max_tokens": 128000, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": true, + "options": {} + }, { "id": "gpt-5.1", "name": "GPT 5.1", @@ -1880,6 +1899,25 @@ "supports_attachments": true, "options": {} }, + { + "id": "gpt-5.2-pro", + "name": "GPT 5.2 Pro", + "cost_per_1m_in": 21, + "cost_per_1m_out": 168, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 2.1, + "context_window": 400000, + "default_max_tokens": 128000, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": true, + "options": {} + }, { "id": "gpt-5.3-chat-latest", "name": "GPT 5.3 Chat", @@ -1969,6 +2007,25 @@ "supports_attachments": true, "options": {} }, + { + "id": "gpt-5.4-pro", + "name": "GPT 5.4 Pro", + "cost_per_1m_in": 30, + "cost_per_1m_out": 180, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 30, + "context_window": 1050000, + "default_max_tokens": 128000, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": true, + "options": {} + }, { "id": "gpt-oss-120b", "name": "gpt-oss-120b", @@ -2466,6 +2523,25 @@ "supports_attachments": true, "options": {} }, + { + "id": "o3-pro", + "name": "O3 Pro", + "cost_per_1m_in": 20, + "cost_per_1m_out": 80, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 20, + "context_window": 200000, + "default_max_tokens": 100000, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": true, + "options": {} + }, { "id": "o4-mini", "name": "O4 Mini", diff --git a/internal/providers/configs/openrouter.json b/internal/providers/configs/openrouter.json index 74f194e69009c327711da0d4050b260beb63779a..39f470de09bb2c3f748fd921c9801d6171f751d0 100644 --- a/internal/providers/configs/openrouter.json +++ b/internal/providers/configs/openrouter.json @@ -560,12 +560,12 @@ { "id": "deepseek/deepseek-chat-v3.1", "name": "DeepSeek: DeepSeek V3.1", - "cost_per_1m_in": 0.2, - "cost_per_1m_out": 0.8, + "cost_per_1m_in": 0.56, + "cost_per_1m_out": 1.68, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.1, + "cost_per_1m_out_cached": 0.28, "context_window": 163840, - "default_max_tokens": 32768, + "default_max_tokens": 16384, "can_reason": true, "reasoning_levels": [ "low", @@ -579,12 +579,12 @@ { "id": "deepseek/deepseek-v3.1-terminus", "name": "DeepSeek: DeepSeek V3.1 Terminus", - "cost_per_1m_in": 0.21, - "cost_per_1m_out": 0.79, + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 0.95, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.13, - "context_window": 163840, - "default_max_tokens": 16384, + "context_window": 131072, + "default_max_tokens": 32768, "can_reason": true, "reasoning_levels": [ "low", @@ -620,9 +620,9 @@ "cost_per_1m_in": 0.27, "cost_per_1m_out": 0.41, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.27, + "cost_per_1m_out_cached": 0, "context_window": 163840, - "default_max_tokens": 81920, + "default_max_tokens": 32768, "can_reason": true, "reasoning_levels": [ "low", @@ -1070,12 +1070,12 @@ { "id": "meta-llama/llama-3.3-70b-instruct", "name": "Meta: Llama 3.3 70B Instruct", - "cost_per_1m_in": 0.13, - "cost_per_1m_out": 0.4, + "cost_per_1m_in": 0.88, + "cost_per_1m_out": 0.88, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 64000, + "default_max_tokens": 1024, "can_reason": false, "supports_attachments": false, "options": {} @@ -1087,8 +1087,8 @@ "cost_per_1m_out": 0, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 65536, - "default_max_tokens": 6553, + "context_window": 128000, + "default_max_tokens": 64000, "can_reason": false, "supports_attachments": false, "options": {} @@ -1122,12 +1122,12 @@ { "id": "minimax/minimax-m2", "name": "MiniMax: MiniMax M2", - "cost_per_1m_in": 0.3, - "cost_per_1m_out": 1.2, + "cost_per_1m_in": 0.255, + "cost_per_1m_out": 1, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.03, - "context_window": 204800, - "default_max_tokens": 65536, + "context_window": 196608, + "default_max_tokens": 98304, "can_reason": true, "reasoning_levels": [ "low", @@ -1294,7 +1294,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 26214, "can_reason": false, "supports_attachments": true, "options": {} @@ -1406,12 +1406,12 @@ { "id": "mistralai/mistral-small-3.2-24b-instruct", "name": "Mistral: Mistral Small 3.2 24B", - "cost_per_1m_in": 0.06, - "cost_per_1m_out": 0.18, + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.3, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.03, + "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 65536, + "default_max_tokens": 13107, "can_reason": false, "supports_attachments": true, "options": {} @@ -1497,12 +1497,12 @@ { "id": "moonshotai/kimi-k2", "name": "MoonshotAI: Kimi K2 0711", - "cost_per_1m_in": 0.57, - "cost_per_1m_out": 2.3, + "cost_per_1m_in": 0.6, + "cost_per_1m_out": 2.5, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.15, "context_window": 131072, - "default_max_tokens": 65536, + "default_max_tokens": 13107, "can_reason": false, "supports_attachments": false, "options": {} @@ -1510,12 +1510,12 @@ { "id": "moonshotai/kimi-k2-0905", "name": "MoonshotAI: Kimi K2 0905", - "cost_per_1m_in": 1, - "cost_per_1m_out": 3, + "cost_per_1m_in": 0.6, + "cost_per_1m_out": 2.5, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.5, + "cost_per_1m_out_cached": 0.6, "context_window": 262144, - "default_max_tokens": 8192, + "default_max_tokens": 16384, "can_reason": false, "supports_attachments": false, "options": {} @@ -1523,12 +1523,12 @@ { "id": "moonshotai/kimi-k2-thinking", "name": "MoonshotAI: Kimi K2 Thinking", - "cost_per_1m_in": 0.47, - "cost_per_1m_out": 2, + "cost_per_1m_in": 0.6, + "cost_per_1m_out": 2.5, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.141, - "context_window": 131072, - "default_max_tokens": 13107, + "cost_per_1m_out_cached": 0.15, + "context_window": 262144, + "default_max_tokens": 131072, "can_reason": true, "reasoning_levels": [ "low", @@ -1543,11 +1543,11 @@ "id": "moonshotai/kimi-k2.5", "name": "MoonshotAI: Kimi K2.5", "cost_per_1m_in": 0.45, - "cost_per_1m_out": 2.2, + "cost_per_1m_out": 2.25, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.225, + "cost_per_1m_out_cached": 0.07, "context_window": 262144, - "default_max_tokens": 32767, + "default_max_tokens": 131072, "can_reason": true, "reasoning_levels": [ "low", @@ -1788,9 +1788,9 @@ "cost_per_1m_in": 0.1, "cost_per_1m_out": 0.4, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.025, + "cost_per_1m_out_cached": 0.03, "context_window": 1047576, - "default_max_tokens": 16384, + "default_max_tokens": 104757, "can_reason": false, "supports_attachments": true, "options": {} @@ -1801,7 +1801,7 @@ "cost_per_1m_in": 2.5, "cost_per_1m_out": 10, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 1.25, + "cost_per_1m_out_cached": 0, "context_window": 128000, "default_max_tokens": 8192, "can_reason": false, @@ -2038,7 +2038,7 @@ "cost_per_1m_in": 1.25, "cost_per_1m_out": 10, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.125, + "cost_per_1m_out_cached": 0.13, "context_window": 400000, "default_max_tokens": 64000, "can_reason": true, @@ -2148,7 +2148,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.175, "context_window": 128000, - "default_max_tokens": 8192, + "default_max_tokens": 16000, "can_reason": false, "supports_attachments": true, "options": {} @@ -2302,8 +2302,8 @@ { "id": "openai/gpt-oss-20b", "name": "OpenAI: gpt-oss-20b", - "cost_per_1m_in": 0.03, - "cost_per_1m_out": 0.14, + "cost_per_1m_in": 0.05, + "cost_per_1m_out": 0.2, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, @@ -2576,8 +2576,8 @@ { "id": "qwen/qwen-vl-max", "name": "Qwen: Qwen VL Max", - "cost_per_1m_in": 0.8, - "cost_per_1m_out": 3.2, + "cost_per_1m_in": 0.52, + "cost_per_1m_out": 2.08, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, @@ -2602,10 +2602,10 @@ { "id": "qwen/qwen-plus", "name": "Qwen: Qwen-Plus", - "cost_per_1m_in": 0.4, - "cost_per_1m_out": 1.2, + "cost_per_1m_in": 0.26, + "cost_per_1m_out": 0.78, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.08, + "cost_per_1m_out_cached": 0.052, "context_window": 1000000, "default_max_tokens": 16384, "can_reason": false, @@ -2638,6 +2638,25 @@ "supports_attachments": false, "options": {} }, + { + "id": "qwen/qwen3-14b", + "name": "Qwen: Qwen3 14B", + "cost_per_1m_in": 0.12, + "cost_per_1m_out": 0.24, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 40960, + "default_max_tokens": 20480, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false, + "options": {} + }, { "id": "qwen/qwen3-235b-a22b", "name": "Qwen: Qwen3 235B A22B", @@ -2673,12 +2692,12 @@ { "id": "qwen/qwen3-235b-a22b-thinking-2507", "name": "Qwen: Qwen3 235B A22B Thinking 2507", - "cost_per_1m_in": 0.11, - "cost_per_1m_out": 0.6, + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 3, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.055, - "context_window": 262144, - "default_max_tokens": 131072, + "cost_per_1m_out_cached": 0, + "context_window": 131072, + "default_max_tokens": 16384, "can_reason": true, "reasoning_levels": [ "low", @@ -2724,11 +2743,11 @@ { "id": "qwen/qwen3-30b-a3b-thinking-2507", "name": "Qwen: Qwen3 30B A3B Thinking 2507", - "cost_per_1m_in": 0.09, - "cost_per_1m_out": 0.3, + "cost_per_1m_in": 0.08, + "cost_per_1m_out": 0.4, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 262144, + "cost_per_1m_out_cached": 0.08, + "context_window": 131072, "default_max_tokens": 65536, "can_reason": true, "reasoning_levels": [ @@ -2813,12 +2832,12 @@ { "id": "qwen/qwen3-coder", "name": "Qwen: Qwen3 Coder 480B A35B", - "cost_per_1m_in": 0.22, + "cost_per_1m_in": 0.4, "cost_per_1m_out": 1.8, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 32768, + "default_max_tokens": 131072, "can_reason": false, "supports_attachments": false, "options": {} @@ -2878,10 +2897,10 @@ { "id": "qwen/qwen3-max", "name": "Qwen: Qwen3 Max", - "cost_per_1m_in": 1.2, - "cost_per_1m_out": 6, + "cost_per_1m_in": 0.78, + "cost_per_1m_out": 3.9, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.24, + "cost_per_1m_out_cached": 0.156, "context_window": 262144, "default_max_tokens": 16384, "can_reason": false, @@ -2910,8 +2929,8 @@ { "id": "qwen/qwen3-next-80b-a3b-instruct", "name": "Qwen: Qwen3 Next 80B A3B Instruct", - "cost_per_1m_in": 0.1, - "cost_per_1m_out": 1.1, + "cost_per_1m_in": 0.15, + "cost_per_1m_out": 1.2, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, @@ -2955,12 +2974,12 @@ { "id": "qwen/qwen3-vl-235b-a22b-instruct", "name": "Qwen: Qwen3 VL 235B A22B Instruct", - "cost_per_1m_in": 0.25, - "cost_per_1m_out": 1.5, + "cost_per_1m_in": 0.26, + "cost_per_1m_out": 1.04, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 256000, - "default_max_tokens": 25600, + "context_window": 131072, + "default_max_tokens": 16384, "can_reason": false, "supports_attachments": true, "options": {} @@ -3069,7 +3088,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 32768, + "default_max_tokens": 131072, "can_reason": true, "reasoning_levels": [ "low", @@ -3102,10 +3121,10 @@ { "id": "qwen/qwen3.5-122b-a10b", "name": "Qwen: Qwen3.5-122B-A10B", - "cost_per_1m_in": 0.3, - "cost_per_1m_out": 2.4, + "cost_per_1m_in": 0.4, + "cost_per_1m_out": 3.2, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.3, + "cost_per_1m_out_cached": 0, "context_window": 262144, "default_max_tokens": 32768, "can_reason": true, @@ -3121,10 +3140,10 @@ { "id": "qwen/qwen3.5-27b", "name": "Qwen: Qwen3.5-27B", - "cost_per_1m_in": 0.27, - "cost_per_1m_out": 2.16, + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 2.4, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.27, + "cost_per_1m_out_cached": 0, "context_window": 262144, "default_max_tokens": 32768, "can_reason": true, @@ -3178,8 +3197,8 @@ { "id": "qwen/qwen3.5-flash-02-23", "name": "Qwen: Qwen3.5-Flash", - "cost_per_1m_in": 0.1, - "cost_per_1m_out": 0.4, + "cost_per_1m_in": 0.065, + "cost_per_1m_out": 0.26, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 1000000, @@ -3368,7 +3387,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.11, "context_window": 131072, - "default_max_tokens": 48000, + "default_max_tokens": 49152, "can_reason": true, "reasoning_levels": [ "low", @@ -3477,12 +3496,12 @@ { "id": "z-ai/glm-4.7", "name": "Z.ai: GLM 4.7", - "cost_per_1m_in": 0.45, - "cost_per_1m_out": 2.2, + "cost_per_1m_in": 0.54, + "cost_per_1m_out": 1.98, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.11, + "cost_per_1m_out_cached": 0.099, "context_window": 204800, - "default_max_tokens": 102400, + "default_max_tokens": 65536, "can_reason": true, "reasoning_levels": [ "low", diff --git a/internal/providers/configs/vercel.json b/internal/providers/configs/vercel.json index df7b99aef6f3ac429cefba92efe52ea8f049b59f..596d3a8896392444556bb9238efbb584a709372f 100644 --- a/internal/providers/configs/vercel.json +++ b/internal/providers/configs/vercel.json @@ -538,6 +538,25 @@ "supports_attachments": false, "options": {} }, + { + "id": "zai/glm-5-turbo", + "name": "GLM 5 Turbo", + "cost_per_1m_in": 1.2, + "cost_per_1m_out": 4, + "cost_per_1m_in_cached": 0.24, + "cost_per_1m_out_cached": 0, + "context_window": 202800, + "default_max_tokens": 8000, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false, + "options": {} + }, { "id": "zai/glm-4.5", "name": "GLM-4.5",