From abdd974daf1ff98af63b5b84318dbe86f5dc1acf Mon Sep 17 00:00:00 2001 From: Charm <124303983+charmcli@users.noreply.github.com> Date: Tue, 12 May 2026 03:14:09 +0000 Subject: [PATCH] chore: auto-update generated files --- internal/providers/configs/aihubmix.json | 18 -- internal/providers/configs/huggingface.json | 12 -- internal/providers/configs/ionet.json | 8 +- internal/providers/configs/opencode-zen.json | 18 ++ internal/providers/configs/openrouter.json | 210 ++++++++++--------- 5 files changed, 136 insertions(+), 130 deletions(-) diff --git a/internal/providers/configs/aihubmix.json b/internal/providers/configs/aihubmix.json index 1e9a977f76e70a3f4033c9ce7cca765f681066d2..c3b0047d19c9eb7bf594e0b759de618b0a17d231 100644 --- a/internal/providers/configs/aihubmix.json +++ b/internal/providers/configs/aihubmix.json @@ -763,24 +763,6 @@ "default_reasoning_effort": "medium", "supports_attachments": false }, - { - "id": "deepseek-v3.2-fast", - "name": "DeepSeek V3.2 Fast", - "cost_per_1m_in": 1.096, - "cost_per_1m_out": 3.288, - "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 1.096, - "context_window": 128000, - "default_max_tokens": 12800, - "can_reason": true, - "reasoning_levels": [ - "low", - "medium", - "high" - ], - "default_reasoning_effort": "medium", - "supports_attachments": false - }, { "id": "deepseek-v3.2-speciale", "name": "DeepSeek V3.2 Speciale", diff --git a/internal/providers/configs/huggingface.json b/internal/providers/configs/huggingface.json index 2622bcf3b64402a3c03bc69b0913a0bd44318903..ed062a6ca3e3816febda46c9d9d94913a7ff6215 100644 --- a/internal/providers/configs/huggingface.json +++ b/internal/providers/configs/huggingface.json @@ -31,18 +31,6 @@ "can_reason": false, "supports_attachments": false }, - { - "id": "Qwen/Qwen3-235B-A22B-Instruct-2507:cerebras", - "name": "Qwen/Qwen3-235B-A22B-Instruct-2507 (cerebras)", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 1.2, - "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 131072, - "default_max_tokens": 8192, - "can_reason": false, - "supports_attachments": false - }, { "id": "Qwen/Qwen3-32B:groq", "name": "Qwen/Qwen3-32B (groq)", diff --git a/internal/providers/configs/ionet.json b/internal/providers/configs/ionet.json index bd3cdd24516f20904795d532423d8ca669437ce2..f99156d2067ef40e7260905af7fdb4f7945d9075 100644 --- a/internal/providers/configs/ionet.json +++ b/internal/providers/configs/ionet.json @@ -166,12 +166,12 @@ { "id": "moonshotai/Kimi-K2.6", "name": "MoonshotAI: Kimi K2.6", - "cost_per_1m_in": 0.8, - "cost_per_1m_out": 4, + "cost_per_1m_in": 0.74, + "cost_per_1m_out": 3.5, "cost_per_1m_in_cached": 0.25, "cost_per_1m_out_cached": 0, - "context_window": 262142, - "default_max_tokens": 26214, + "context_window": 32768, + "default_max_tokens": 3276, "can_reason": true, "supports_attachments": true }, diff --git a/internal/providers/configs/opencode-zen.json b/internal/providers/configs/opencode-zen.json index 61887671a4de8b1df77df9cf83a209bd4f90900a..2b1f778498b3b6f120b74c8faa89bc4dd1e6e04a 100644 --- a/internal/providers/configs/opencode-zen.json +++ b/internal/providers/configs/opencode-zen.json @@ -169,6 +169,24 @@ "default_reasoning_effort": "medium", "supports_attachments": true }, + { + "id": "deepseek-v4-flash-free", + "name": "DeepSeek V4 Flash Free", + "cost_per_1m_in": 0, + "cost_per_1m_out": 0, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 1000000, + "default_max_tokens": 384000, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false + }, { "id": "glm-5", "name": "GLM-5", diff --git a/internal/providers/configs/openrouter.json b/internal/providers/configs/openrouter.json index c4add838f8991f562a42c24337d96565f20c9e00..d93c619323b2d1387c2480771959e228743b5cbc 100644 --- a/internal/providers/configs/openrouter.json +++ b/internal/providers/configs/openrouter.json @@ -319,6 +319,24 @@ "default_reasoning_effort": "medium", "supports_attachments": false }, + { + "id": "arcee-ai/trinity-large-thinking:free", + "name": "Arcee AI: Trinity Large Thinking (free)", + "cost_per_1m_in": 0, + "cost_per_1m_out": 0, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 262144, + "default_max_tokens": 40000, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false + }, { "id": "arcee-ai/trinity-mini", "name": "Arcee AI: Trinity Mini", @@ -496,12 +514,12 @@ { "id": "deepseek/deepseek-chat", "name": "DeepSeek: DeepSeek V3", - "cost_per_1m_in": 0.32, - "cost_per_1m_out": 0.89, + "cost_per_1m_in": 0.4, + "cost_per_1m_out": 1.3, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 163840, - "default_max_tokens": 8192, + "context_window": 64000, + "default_max_tokens": 8000, "can_reason": false, "supports_attachments": false }, @@ -520,10 +538,10 @@ { "id": "deepseek/deepseek-chat-v3.1", "name": "DeepSeek: DeepSeek V3.1", - "cost_per_1m_in": 0.56, - "cost_per_1m_out": 1.68, + "cost_per_1m_in": 0.21, + "cost_per_1m_out": 0.79, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.28, + "cost_per_1m_out_cached": 0.13, "context_window": 163840, "default_max_tokens": 16384, "can_reason": true, @@ -556,10 +574,10 @@ { "id": "deepseek/deepseek-v3.2", "name": "DeepSeek: DeepSeek V3.2", - "cost_per_1m_in": 0.5, - "cost_per_1m_out": 1.5, + "cost_per_1m_in": 0.26, + "cost_per_1m_out": 0.38, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.25, + "cost_per_1m_out_cached": 0.13, "context_window": 163840, "default_max_tokens": 81920, "can_reason": true, @@ -922,12 +940,12 @@ { "id": "google/gemma-4-26b-a4b-it", "name": "Google: Gemma 4 26B A4B ", - "cost_per_1m_in": 0.07, - "cost_per_1m_out": 0.34, + "cost_per_1m_in": 0.15, + "cost_per_1m_out": 0.6, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 8192, + "default_max_tokens": 131072, "can_reason": true, "reasoning_levels": [ "low", @@ -1048,12 +1066,12 @@ { "id": "meta-llama/llama-3.3-70b-instruct", "name": "Meta: Llama 3.3 70B Instruct", - "cost_per_1m_in": 0.59, - "cost_per_1m_out": 0.79, + "cost_per_1m_in": 0.135, + "cost_per_1m_out": 0.4, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 16384, + "default_max_tokens": 60000, "can_reason": false, "supports_attachments": false }, @@ -1156,8 +1174,8 @@ { "id": "minimax/minimax-m2.7", "name": "MiniMax: MiniMax M2.7", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 2.4, + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 1.2, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.06, "context_window": 204800, @@ -1450,12 +1468,12 @@ { "id": "moonshotai/kimi-k2-0905", "name": "MoonshotAI: Kimi K2 0905", - "cost_per_1m_in": 1, - "cost_per_1m_out": 3, + "cost_per_1m_in": 0.6, + "cost_per_1m_out": 2.5, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.5, + "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 8192, + "default_max_tokens": 131072, "can_reason": false, "supports_attachments": false }, @@ -1481,11 +1499,11 @@ "id": "moonshotai/kimi-k2.5", "name": "MoonshotAI: Kimi K2.5", "cost_per_1m_in": 0.45, - "cost_per_1m_out": 2.25, + "cost_per_1m_out": 2, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.07, + "cost_per_1m_out_cached": 0.1, "context_window": 262144, - "default_max_tokens": 32000, + "default_max_tokens": 131072, "can_reason": true, "reasoning_levels": [ "low", @@ -1767,7 +1785,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.5, "context_window": 1047576, - "default_max_tokens": 104757, + "default_max_tokens": 16384, "can_reason": false, "supports_attachments": true }, @@ -1801,7 +1819,7 @@ "cost_per_1m_in": 2.5, "cost_per_1m_out": 10, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 1.25, + "cost_per_1m_out_cached": 0, "context_window": 128000, "default_max_tokens": 8192, "can_reason": false, @@ -2668,12 +2686,12 @@ { "id": "qwen/qwen3-14b", "name": "Qwen: Qwen3 14B", - "cost_per_1m_in": 0.12, - "cost_per_1m_out": 0.24, + "cost_per_1m_in": 0.2275, + "cost_per_1m_out": 0.91, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 40960, - "default_max_tokens": 8192, + "context_window": 131072, + "default_max_tokens": 4096, "can_reason": true, "reasoning_levels": [ "low", @@ -2704,10 +2722,10 @@ { "id": "qwen/qwen3-235b-a22b-2507", "name": "Qwen: Qwen3 235B A22B Instruct 2507", - "cost_per_1m_in": 0.1, - "cost_per_1m_out": 0.1, + "cost_per_1m_in": 0.2, + "cost_per_1m_out": 0.8, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.1, + "cost_per_1m_out_cached": 0, "context_window": 262144, "default_max_tokens": 131072, "can_reason": false, @@ -2734,12 +2752,12 @@ { "id": "qwen/qwen3-30b-a3b", "name": "Qwen: Qwen3 30B A3B", - "cost_per_1m_in": 0.09, - "cost_per_1m_out": 0.45, + "cost_per_1m_in": 0.13, + "cost_per_1m_out": 0.52, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 40960, - "default_max_tokens": 10000, + "context_window": 131072, + "default_max_tokens": 4096, "can_reason": true, "reasoning_levels": [ "low", @@ -2755,9 +2773,9 @@ "cost_per_1m_in": 0.1, "cost_per_1m_out": 0.3, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.1, + "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 26214, "can_reason": false, "supports_attachments": false }, @@ -2866,7 +2884,7 @@ { "id": "qwen/qwen3-coder-next", "name": "Qwen: Qwen3 Coder Next", - "cost_per_1m_in": 0.12, + "cost_per_1m_in": 0.11, "cost_per_1m_out": 0.8, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.07, @@ -2920,12 +2938,12 @@ { "id": "qwen/qwen3-next-80b-a3b-instruct", "name": "Qwen: Qwen3 Next 80B A3B Instruct", - "cost_per_1m_in": 0.1, + "cost_per_1m_in": 0.09, "cost_per_1m_out": 1.1, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.07, + "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 8192, "can_reason": false, "supports_attachments": false }, @@ -2962,12 +2980,12 @@ { "id": "qwen/qwen3-vl-235b-a22b-instruct", "name": "Qwen: Qwen3 VL 235B A22B Instruct", - "cost_per_1m_in": 0.2, - "cost_per_1m_out": 0.88, + "cost_per_1m_in": 0.26, + "cost_per_1m_out": 1.04, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.11, - "context_window": 262144, - "default_max_tokens": 8192, + "cost_per_1m_out_cached": 0, + "context_window": 131072, + "default_max_tokens": 16384, "can_reason": false, "supports_attachments": true }, @@ -3034,12 +3052,12 @@ { "id": "qwen/qwen3-vl-8b-instruct", "name": "Qwen: Qwen3 VL 8B Instruct", - "cost_per_1m_in": 0.117, - "cost_per_1m_out": 0.455, + "cost_per_1m_in": 0.25, + "cost_per_1m_out": 0.75, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 131072, - "default_max_tokens": 16384, + "cost_per_1m_out_cached": 0.12, + "context_window": 262144, + "default_max_tokens": 131072, "can_reason": false, "supports_attachments": true }, @@ -3118,10 +3136,10 @@ { "id": "qwen/qwen3.5-122b-a10b", "name": "Qwen: Qwen3.5-122B-A10B", - "cost_per_1m_in": 0.26, - "cost_per_1m_out": 2.08, + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 2.4, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.3, "context_window": 262144, "default_max_tokens": 32768, "can_reason": true, @@ -3136,12 +3154,12 @@ { "id": "qwen/qwen3.5-27b", "name": "Qwen: Qwen3.5-27B", - "cost_per_1m_in": 0.3, - "cost_per_1m_out": 2.4, + "cost_per_1m_in": 0.26, + "cost_per_1m_out": 2.6, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 32768, + "default_max_tokens": 40960, "can_reason": true, "reasoning_levels": [ "low", @@ -3154,10 +3172,10 @@ { "id": "qwen/qwen3.5-35b-a3b", "name": "Qwen: Qwen3.5-35B-A3B", - "cost_per_1m_in": 0.15, - "cost_per_1m_out": 1, + "cost_per_1m_in": 0.16, + "cost_per_1m_out": 1.2, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.05, + "cost_per_1m_out_cached": 0, "context_window": 262144, "default_max_tokens": 131072, "can_reason": true, @@ -3172,12 +3190,12 @@ { "id": "qwen/qwen3.5-9b", "name": "Qwen: Qwen3.5-9B", - "cost_per_1m_in": 0.04, + "cost_per_1m_in": 0.1, "cost_per_1m_out": 0.15, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 40960, + "default_max_tokens": 26214, "can_reason": true, "reasoning_levels": [ "low", @@ -3208,12 +3226,12 @@ { "id": "qwen/qwen3.6-27b", "name": "Qwen: Qwen3.6 27B", - "cost_per_1m_in": 0.32, - "cost_per_1m_out": 3.2, + "cost_per_1m_in": 0.5, + "cost_per_1m_out": 2, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.25, "context_window": 262144, - "default_max_tokens": 40960, + "default_max_tokens": 32768, "can_reason": true, "reasoning_levels": [ "low", @@ -3226,10 +3244,10 @@ { "id": "qwen/qwen3.6-35b-a3b", "name": "Qwen: Qwen3.6 35B A3B", - "cost_per_1m_in": 0.17, - "cost_per_1m_out": 1.2, + "cost_per_1m_in": 0.15, + "cost_per_1m_out": 1, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.05, "context_window": 262144, "default_max_tokens": 131072, "can_reason": true, @@ -3313,9 +3331,9 @@ "cost_per_1m_in": 0.1, "cost_per_1m_out": 0.3, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.03, + "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 32768, "can_reason": true, "reasoning_levels": [ "low", @@ -3513,7 +3531,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.11, "context_window": 131072, - "default_max_tokens": 48000, + "default_max_tokens": 49152, "can_reason": true, "reasoning_levels": [ "low", @@ -3616,12 +3634,12 @@ { "id": "z-ai/glm-4.7", "name": "Z.ai: GLM 4.7", - "cost_per_1m_in": 0.52, - "cost_per_1m_out": 1.85, + "cost_per_1m_in": 0.45, + "cost_per_1m_out": 2.2, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.12, - "context_window": 202752, - "default_max_tokens": 101376, + "cost_per_1m_out_cached": 0.11, + "context_window": 204800, + "default_max_tokens": 102400, "can_reason": true, "reasoning_levels": [ "low", @@ -3634,12 +3652,12 @@ { "id": "z-ai/glm-4.7-flash", "name": "Z.ai: GLM 4.7 Flash", - "cost_per_1m_in": 0.06, - "cost_per_1m_out": 0.4, + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.43, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.01, + "cost_per_1m_out_cached": 0, "context_window": 202752, - "default_max_tokens": 8192, + "default_max_tokens": 101376, "can_reason": true, "reasoning_levels": [ "low", @@ -3766,10 +3784,10 @@ { "id": "x-ai/grok-3", "name": "xAI: Grok 3", - "cost_per_1m_in": 3, - "cost_per_1m_out": 15, + "cost_per_1m_in": 5, + "cost_per_1m_out": 25, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.75, + "cost_per_1m_out_cached": 1.25, "context_window": 131072, "default_max_tokens": 13107, "can_reason": false, @@ -3778,10 +3796,10 @@ { "id": "x-ai/grok-3-beta", "name": "xAI: Grok 3 Beta", - "cost_per_1m_in": 3, - "cost_per_1m_out": 15, + "cost_per_1m_in": 5, + "cost_per_1m_out": 25, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.75, + "cost_per_1m_out_cached": 1.25, "context_window": 131072, "default_max_tokens": 13107, "can_reason": false, @@ -3790,10 +3808,10 @@ { "id": "x-ai/grok-3-mini", "name": "xAI: Grok 3 Mini", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 4, + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 0.5, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.15, + "cost_per_1m_out_cached": 0.075, "context_window": 131072, "default_max_tokens": 13107, "can_reason": true, @@ -3808,10 +3826,10 @@ { "id": "x-ai/grok-3-mini-beta", "name": "xAI: Grok 3 Mini Beta", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 4, + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 0.5, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.15, + "cost_per_1m_out_cached": 0.075, "context_window": 131072, "default_max_tokens": 13107, "can_reason": true,