From 139b0f1108e58c87cef3e39b2079be00fa8073f5 Mon Sep 17 00:00:00 2001 From: Charm <124303983+charmcli@users.noreply.github.com> Date: Mon, 16 Mar 2026 03:10:51 +0000 Subject: [PATCH] chore: auto-update generated files --- internal/providers/configs/aihubmix.json | 38 ++++ internal/providers/configs/huggingface.json | 13 ++ internal/providers/configs/openrouter.json | 225 +++++++++++--------- internal/providers/configs/venice.json | 2 +- 4 files changed, 171 insertions(+), 107 deletions(-) diff --git a/internal/providers/configs/aihubmix.json b/internal/providers/configs/aihubmix.json index 0fb3865cbe5848c67cea87773944927961085b30..b90f951fb38ab4525b2c3496a078fc0425e70a09 100644 --- a/internal/providers/configs/aihubmix.json +++ b/internal/providers/configs/aihubmix.json @@ -1425,6 +1425,25 @@ "supports_attachments": false, "options": {} }, + { + "id": "glm-5-turbo", + "name": "GLM 5 Turbo", + "cost_per_1m_in": 1.2, + "cost_per_1m_out": 3.9996, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0.24, + "context_window": 202752, + "default_max_tokens": 20275, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false, + "options": {} + }, { "id": "gpt-4.1", "name": "GPT 4.1", @@ -2109,6 +2128,25 @@ "supports_attachments": true, "options": {} }, + { + "id": "grok-4.20-multi-agent-beta-0309", + "name": "Grok 4.20 Multi Agent Beta 0309", + "cost_per_1m_in": 2, + "cost_per_1m_out": 6, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0.2, + "context_window": 2000000, + "default_max_tokens": 200000, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": true, + "options": {} + }, { "id": "grok-code-fast-1", "name": "Grok Code Fast 1", diff --git a/internal/providers/configs/huggingface.json b/internal/providers/configs/huggingface.json index fe61e6320d6c8ba76df8e6c547b0ec0e800f1e32..8cb6ca8360a6538f095648587892e1c1965ec738 100644 --- a/internal/providers/configs/huggingface.json +++ b/internal/providers/configs/huggingface.json @@ -7,6 +7,19 @@ "default_large_model_id": "moonshotai/Kimi-K2-Instruct-0905:groq", "default_small_model_id": "openai/gpt-oss-20b:groq", "models": [ + { + "id": "MiniMaxAI/MiniMax-M2.1:fireworks-ai", + "name": "MiniMaxAI/MiniMax-M2.1 (fireworks-ai)", + "cost_per_1m_in": 0, + "cost_per_1m_out": 0, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 204800, + "default_max_tokens": 8192, + "can_reason": false, + "supports_attachments": false, + "options": {} + }, { "id": "MiniMaxAI/MiniMax-M2.5:fireworks-ai", "name": "MiniMaxAI/MiniMax-M2.5 (fireworks-ai)", diff --git a/internal/providers/configs/openrouter.json b/internal/providers/configs/openrouter.json index 89770d8c5631350e703b69431bc60ca269ef87ca..74f194e69009c327711da0d4050b260beb63779a 100644 --- a/internal/providers/configs/openrouter.json +++ b/internal/providers/configs/openrouter.json @@ -547,31 +547,25 @@ { "id": "deepseek/deepseek-chat-v3-0324", "name": "DeepSeek: DeepSeek V3 0324", - "cost_per_1m_in": 0.77, - "cost_per_1m_out": 0.77, + "cost_per_1m_in": 0.27, + "cost_per_1m_out": 1.12, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.135, "context_window": 163840, - "default_max_tokens": 65536, - "can_reason": true, - "reasoning_levels": [ - "low", - "medium", - "high" - ], - "default_reasoning_effort": "medium", + "default_max_tokens": 81920, + "can_reason": false, "supports_attachments": false, "options": {} }, { "id": "deepseek/deepseek-chat-v3.1", "name": "DeepSeek: DeepSeek V3.1", - "cost_per_1m_in": 0.21, - "cost_per_1m_out": 0.79, + "cost_per_1m_in": 0.2, + "cost_per_1m_out": 0.8, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.13, + "cost_per_1m_out_cached": 0.1, "context_window": 163840, - "default_max_tokens": 16384, + "default_max_tokens": 32768, "can_reason": true, "reasoning_levels": [ "low", @@ -585,12 +579,12 @@ { "id": "deepseek/deepseek-v3.1-terminus", "name": "DeepSeek: DeepSeek V3.1 Terminus", - "cost_per_1m_in": 0.23, - "cost_per_1m_out": 0.9, + "cost_per_1m_in": 0.21, + "cost_per_1m_out": 0.79, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.115, + "cost_per_1m_out_cached": 0.13, "context_window": 163840, - "default_max_tokens": 32768, + "default_max_tokens": 16384, "can_reason": true, "reasoning_levels": [ "low", @@ -604,12 +598,12 @@ { "id": "deepseek/deepseek-v3.2", "name": "DeepSeek: DeepSeek V3.2", - "cost_per_1m_in": 0.56, - "cost_per_1m_out": 1.68, + "cost_per_1m_in": 0.26, + "cost_per_1m_out": 0.38, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.13, "context_window": 163840, - "default_max_tokens": 32768, + "default_max_tokens": 81920, "can_reason": true, "reasoning_levels": [ "low", @@ -626,9 +620,9 @@ "cost_per_1m_in": 0.27, "cost_per_1m_out": 0.41, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.27, "context_window": 163840, - "default_max_tokens": 32768, + "default_max_tokens": 81920, "can_reason": true, "reasoning_levels": [ "low", @@ -986,7 +980,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, + "supports_attachments": true, "options": {} }, { @@ -1076,12 +1070,12 @@ { "id": "meta-llama/llama-3.3-70b-instruct", "name": "Meta: Llama 3.3 70B Instruct", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 0.6, + "cost_per_1m_in": 0.13, + "cost_per_1m_out": 0.4, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.6, + "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 65536, + "default_max_tokens": 64000, "can_reason": false, "supports_attachments": false, "options": {} @@ -1093,8 +1087,8 @@ "cost_per_1m_out": 0, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 128000, - "default_max_tokens": 64000, + "context_window": 65536, + "default_max_tokens": 6553, "can_reason": false, "supports_attachments": false, "options": {} @@ -1128,12 +1122,12 @@ { "id": "minimax/minimax-m2", "name": "MiniMax: MiniMax M2", - "cost_per_1m_in": 0.255, - "cost_per_1m_out": 1, + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 1.2, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.03, - "context_window": 196608, - "default_max_tokens": 98304, + "context_window": 204800, + "default_max_tokens": 65536, "can_reason": true, "reasoning_levels": [ "low", @@ -1182,6 +1176,25 @@ "supports_attachments": false, "options": {} }, + { + "id": "minimax/minimax-m2.5:free", + "name": "MiniMax: MiniMax M2.5 (free)", + "cost_per_1m_in": 0, + "cost_per_1m_out": 0, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 196608, + "default_max_tokens": 98304, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false, + "options": {} + }, { "id": "mistralai/mistral-large", "name": "Mistral Large", @@ -1294,7 +1307,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 65536, + "default_max_tokens": 13107, "can_reason": false, "supports_attachments": true, "options": {} @@ -1484,12 +1497,12 @@ { "id": "moonshotai/kimi-k2", "name": "MoonshotAI: Kimi K2 0711", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 2.5, + "cost_per_1m_in": 0.57, + "cost_per_1m_out": 2.3, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.15, + "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 13107, + "default_max_tokens": 65536, "can_reason": false, "supports_attachments": false, "options": {} @@ -1497,12 +1510,12 @@ { "id": "moonshotai/kimi-k2-0905", "name": "MoonshotAI: Kimi K2 0905", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 2.5, + "cost_per_1m_in": 1, + "cost_per_1m_out": 3, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.15, + "cost_per_1m_out_cached": 0.5, "context_window": 262144, - "default_max_tokens": 26214, + "default_max_tokens": 8192, "can_reason": false, "supports_attachments": false, "options": {} @@ -1510,12 +1523,12 @@ { "id": "moonshotai/kimi-k2-thinking", "name": "MoonshotAI: Kimi K2 Thinking", - "cost_per_1m_in": 0.55, - "cost_per_1m_out": 2.5, + "cost_per_1m_in": 0.47, + "cost_per_1m_out": 2, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 262144, - "default_max_tokens": 131072, + "cost_per_1m_out_cached": 0.141, + "context_window": 131072, + "default_max_tokens": 13107, "can_reason": true, "reasoning_levels": [ "low", @@ -1529,12 +1542,12 @@ { "id": "moonshotai/kimi-k2.5", "name": "MoonshotAI: Kimi K2.5", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 2.8, + "cost_per_1m_in": 0.45, + "cost_per_1m_out": 2.2, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.225, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 32767, "can_reason": true, "reasoning_levels": [ "low", @@ -1775,9 +1788,9 @@ "cost_per_1m_in": 0.1, "cost_per_1m_out": 0.4, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.03, + "cost_per_1m_out_cached": 0.025, "context_window": 1047576, - "default_max_tokens": 104757, + "default_max_tokens": 16384, "can_reason": false, "supports_attachments": true, "options": {} @@ -1968,9 +1981,9 @@ "cost_per_1m_in": 0.25, "cost_per_1m_out": 2, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.03, + "cost_per_1m_out_cached": 0.025, "context_window": 400000, - "default_max_tokens": 40000, + "default_max_tokens": 64000, "can_reason": true, "reasoning_levels": [ "low", @@ -1987,9 +2000,9 @@ "cost_per_1m_in": 0.05, "cost_per_1m_out": 0.4, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.01, + "cost_per_1m_out_cached": 0.005, "context_window": 400000, - "default_max_tokens": 40000, + "default_max_tokens": 64000, "can_reason": true, "reasoning_levels": [ "low", @@ -2625,25 +2638,6 @@ "supports_attachments": false, "options": {} }, - { - "id": "qwen/qwen3-14b", - "name": "Qwen: Qwen3 14B", - "cost_per_1m_in": 0.2275, - "cost_per_1m_out": 0.91, - "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 131072, - "default_max_tokens": 4096, - "can_reason": true, - "reasoning_levels": [ - "low", - "medium", - "high" - ], - "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} - }, { "id": "qwen/qwen3-235b-a22b", "name": "Qwen: Qwen3 235B A22B", @@ -2666,8 +2660,8 @@ { "id": "qwen/qwen3-235b-a22b-2507", "name": "Qwen: Qwen3 235B A22B Instruct 2507", - "cost_per_1m_in": 0.1, - "cost_per_1m_out": 0.1, + "cost_per_1m_in": 0.2, + "cost_per_1m_out": 0.8, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, @@ -2730,11 +2724,11 @@ { "id": "qwen/qwen3-30b-a3b-thinking-2507", "name": "Qwen: Qwen3 30B A3B Thinking 2507", - "cost_per_1m_in": 0.08, - "cost_per_1m_out": 0.4, + "cost_per_1m_in": 0.09, + "cost_per_1m_out": 0.3, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.08, - "context_window": 131072, + "cost_per_1m_out_cached": 0, + "context_window": 262144, "default_max_tokens": 65536, "can_reason": true, "reasoning_levels": [ @@ -2806,12 +2800,12 @@ { "id": "qwen/qwen3-coder-30b-a3b-instruct", "name": "Qwen: Qwen3 Coder 30B A3B Instruct", - "cost_per_1m_in": 0.07, - "cost_per_1m_out": 0.28, + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.3, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 26214, "can_reason": false, "supports_attachments": false, "options": {} @@ -2820,11 +2814,11 @@ "id": "qwen/qwen3-coder", "name": "Qwen: Qwen3 Coder 480B A35B", "cost_per_1m_in": 0.22, - "cost_per_1m_out": 1, + "cost_per_1m_out": 1.8, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.022, + "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 26214, + "default_max_tokens": 32768, "can_reason": false, "supports_attachments": false, "options": {} @@ -2858,12 +2852,12 @@ { "id": "qwen/qwen3-coder-next", "name": "Qwen: Qwen3 Coder Next", - "cost_per_1m_in": 0.18, - "cost_per_1m_out": 1.35, + "cost_per_1m_in": 0.12, + "cost_per_1m_out": 0.75, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.18, + "cost_per_1m_out_cached": 0.06, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 32768, "can_reason": false, "supports_attachments": false, "options": {} @@ -3075,7 +3069,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 32768, "can_reason": true, "reasoning_levels": [ "low", @@ -3127,10 +3121,10 @@ { "id": "qwen/qwen3.5-27b", "name": "Qwen: Qwen3.5-27B", - "cost_per_1m_in": 0.195, - "cost_per_1m_out": 1.56, + "cost_per_1m_in": 0.27, + "cost_per_1m_out": 2.16, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.27, "context_window": 262144, "default_max_tokens": 32768, "can_reason": true, @@ -3388,12 +3382,12 @@ { "id": "z-ai/glm-4.5-air", "name": "Z.ai: GLM 4.5 Air", - "cost_per_1m_in": 0.2, - "cost_per_1m_out": 1.2, + "cost_per_1m_in": 0.14, + "cost_per_1m_out": 0.86, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 13107, + "default_max_tokens": 65536, "can_reason": true, "reasoning_levels": [ "low", @@ -3483,12 +3477,12 @@ { "id": "z-ai/glm-4.7", "name": "Z.ai: GLM 4.7", - "cost_per_1m_in": 0.54, - "cost_per_1m_out": 1.98, + "cost_per_1m_in": 0.45, + "cost_per_1m_out": 2.2, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.099, + "cost_per_1m_out_cached": 0.11, "context_window": 204800, - "default_max_tokens": 65536, + "default_max_tokens": 102400, "can_reason": true, "reasoning_levels": [ "low", @@ -3537,6 +3531,25 @@ "supports_attachments": false, "options": {} }, + { + "id": "z-ai/glm-5-turbo", + "name": "Z.ai: GLM 5 Turbo", + "cost_per_1m_in": 0.96, + "cost_per_1m_out": 3.2, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0.192, + "context_window": 202752, + "default_max_tokens": 65536, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false, + "options": {} + }, { "id": "x-ai/grok-3", "name": "xAI: Grok 3", diff --git a/internal/providers/configs/venice.json b/internal/providers/configs/venice.json index 3204aa375ec9b04541d69f61fba874e3371a2657..277e5691e8a7346f6be165a19a1417d03e2865d6 100644 --- a/internal/providers/configs/venice.json +++ b/internal/providers/configs/venice.json @@ -709,7 +709,7 @@ "context_window": 128000, "default_max_tokens": 32000, "can_reason": false, - "supports_attachments": false, + "supports_attachments": true, "options": {} }, {