From 46f41762a0ae35b909150ae36c1d707ddcb065a2 Mon Sep 17 00:00:00 2001 From: Charm <124303983+charmcli@users.noreply.github.com> Date: Sun, 3 May 2026 23:00:05 +0000 Subject: [PATCH] chore: auto-update generated files --- internal/providers/configs/aihubmix.json | 20 +-- internal/providers/configs/huggingface.json | 12 -- internal/providers/configs/openrouter.json | 164 +++++++++----------- 3 files changed, 77 insertions(+), 119 deletions(-) diff --git a/internal/providers/configs/aihubmix.json b/internal/providers/configs/aihubmix.json index dac5362633b3493dac9a7c362340c3a5f569a1cd..d47e8fbe4bfb22f322c432fc25a15abb72712c8c 100644 --- a/internal/providers/configs/aihubmix.json +++ b/internal/providers/configs/aihubmix.json @@ -835,24 +835,6 @@ "default_reasoning_effort": "medium", "supports_attachments": false }, - { - "id": "deepseek-v4-flash-think", - "name": "DeepSeek V4 Flash Thinking", - "cost_per_1m_in": 0.154, - "cost_per_1m_out": 0.308, - "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.00308, - "context_window": 1000000, - "default_max_tokens": 384000, - "can_reason": true, - "reasoning_levels": [ - "low", - "medium", - "high" - ], - "default_reasoning_effort": "medium", - "supports_attachments": false - }, { "id": "deepseek-v4-pro", "name": "DeepSeek V4 Pro", @@ -3275,7 +3257,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true + "supports_attachments": false }, { "id": "qwen3.6-plus", diff --git a/internal/providers/configs/huggingface.json b/internal/providers/configs/huggingface.json index 2622bcf3b64402a3c03bc69b0913a0bd44318903..e263a5b0194439c80f256679784f77c8027ff177 100644 --- a/internal/providers/configs/huggingface.json +++ b/internal/providers/configs/huggingface.json @@ -79,18 +79,6 @@ "can_reason": false, "supports_attachments": false }, - { - "id": "Qwen/Qwen3-VL-30B-A3B-Thinking:fireworks-ai", - "name": "Qwen/Qwen3-VL-30B-A3B-Thinking (fireworks-ai)", - "cost_per_1m_in": 0, - "cost_per_1m_out": 0, - "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 262144, - "default_max_tokens": 8192, - "can_reason": false, - "supports_attachments": false - }, { "id": "deepseek-ai/DeepSeek-V3.2:fireworks-ai", "name": "deepseek-ai/DeepSeek-V3.2 (fireworks-ai)", diff --git a/internal/providers/configs/openrouter.json b/internal/providers/configs/openrouter.json index 7ecd5a664c7de587afee13038361effb4f5a4e7f..9634dbb6867b9d578f76956bcad030b5fcf42994 100644 --- a/internal/providers/configs/openrouter.json +++ b/internal/providers/configs/openrouter.json @@ -568,8 +568,8 @@ { "id": "deepseek/deepseek-v3.1-terminus", "name": "DeepSeek: DeepSeek V3.1 Terminus", - "cost_per_1m_in": 0.21, - "cost_per_1m_out": 0.79, + "cost_per_1m_in": 0.27, + "cost_per_1m_out": 0.95, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.13, "context_window": 163840, @@ -586,12 +586,12 @@ { "id": "deepseek/deepseek-v3.2", "name": "DeepSeek: DeepSeek V3.2", - "cost_per_1m_in": 0.26, - "cost_per_1m_out": 0.38, + "cost_per_1m_in": 0.5, + "cost_per_1m_out": 1.5, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.13, + "cost_per_1m_out_cached": 0.25, "context_window": 163840, - "default_max_tokens": 8192, + "default_max_tokens": 81920, "can_reason": true, "reasoning_levels": [ "low", @@ -640,12 +640,12 @@ { "id": "deepseek/deepseek-v4-pro", "name": "DeepSeek: DeepSeek V4 Pro", - "cost_per_1m_in": 0.435, - "cost_per_1m_out": 0.87, + "cost_per_1m_in": 1.74, + "cost_per_1m_out": 3.48, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.00363, + "cost_per_1m_out_cached": 0.145, "context_window": 1048576, - "default_max_tokens": 192000, + "default_max_tokens": 196608, "can_reason": true, "reasoning_levels": [ "low", @@ -919,27 +919,15 @@ "can_reason": false, "supports_attachments": true }, - { - "id": "google/gemma-3-27b-it", - "name": "Google: Gemma 3 27B", - "cost_per_1m_in": 0.08, - "cost_per_1m_out": 0.16, - "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 131072, - "default_max_tokens": 8192, - "can_reason": false, - "supports_attachments": true - }, { "id": "google/gemma-4-26b-a4b-it", "name": "Google: Gemma 4 26B A4B ", - "cost_per_1m_in": 0.15, - "cost_per_1m_out": 0.6, + "cost_per_1m_in": 0.07, + "cost_per_1m_out": 0.34, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 8192, "can_reason": true, "reasoning_levels": [ "low", @@ -1040,8 +1028,8 @@ "cost_per_1m_out": 1.2, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.06, - "context_window": 262144, - "default_max_tokens": 72000, + "context_window": 256000, + "default_max_tokens": 40000, "can_reason": false, "supports_attachments": false }, @@ -1060,12 +1048,12 @@ { "id": "meta-llama/llama-3.3-70b-instruct", "name": "Meta: Llama 3.3 70B Instruct", - "cost_per_1m_in": 0.13, - "cost_per_1m_out": 0.4, + "cost_per_1m_in": 0.6, + "cost_per_1m_out": 0.6, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.6, "context_window": 131072, - "default_max_tokens": 13107, + "default_max_tokens": 65536, "can_reason": false, "supports_attachments": false }, @@ -1282,12 +1270,12 @@ { "id": "mistralai/ministral-3b-2512", "name": "Mistral: Ministral 3 3B 2512", - "cost_per_1m_in": 0.15, - "cost_per_1m_out": 0.15, + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.1, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.01, "context_window": 131072, - "default_max_tokens": 65536, + "default_max_tokens": 13107, "can_reason": false, "supports_attachments": true }, @@ -1456,12 +1444,12 @@ { "id": "moonshotai/kimi-k2-0905", "name": "MoonshotAI: Kimi K2 0905", - "cost_per_1m_in": 1, - "cost_per_1m_out": 3, + "cost_per_1m_in": 0.6, + "cost_per_1m_out": 2.5, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.5, + "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 8192, + "default_max_tokens": 131072, "can_reason": false, "supports_attachments": false }, @@ -1471,7 +1459,7 @@ "cost_per_1m_in": 0.6, "cost_per_1m_out": 2.5, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.6, + "cost_per_1m_out_cached": 0, "context_window": 262144, "default_max_tokens": 131072, "can_reason": true, @@ -1486,10 +1474,10 @@ { "id": "moonshotai/kimi-k2.5", "name": "MoonshotAI: Kimi K2.5", - "cost_per_1m_in": 0.44, - "cost_per_1m_out": 2.2, + "cost_per_1m_in": 0.49, + "cost_per_1m_out": 2.5, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.22, + "cost_per_1m_out_cached": 0.2, "context_window": 262144, "default_max_tokens": 131072, "can_reason": true, @@ -1504,12 +1492,12 @@ { "id": "moonshotai/kimi-k2.6", "name": "MoonshotAI: Kimi K2.6", - "cost_per_1m_in": 0.95, - "cost_per_1m_out": 4, + "cost_per_1m_in": 0.75, + "cost_per_1m_out": 3.5, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.16, + "cost_per_1m_out_cached": 0.15, "context_window": 262144, - "default_max_tokens": 26214, + "default_max_tokens": 8192, "can_reason": true, "reasoning_levels": [ "low", @@ -1807,7 +1795,7 @@ "cost_per_1m_in": 2.5, "cost_per_1m_out": 10, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 1.25, + "cost_per_1m_out_cached": 0, "context_window": 128000, "default_max_tokens": 8192, "can_reason": false, @@ -2686,12 +2674,12 @@ { "id": "qwen/qwen3-14b", "name": "Qwen: Qwen3 14B", - "cost_per_1m_in": 0.2275, - "cost_per_1m_out": 0.91, + "cost_per_1m_in": 0.12, + "cost_per_1m_out": 0.24, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 131072, - "default_max_tokens": 4096, + "context_window": 40960, + "default_max_tokens": 8192, "can_reason": true, "reasoning_levels": [ "low", @@ -2752,12 +2740,12 @@ { "id": "qwen/qwen3-30b-a3b", "name": "Qwen: Qwen3 30B A3B", - "cost_per_1m_in": 0.08, - "cost_per_1m_out": 0.28, + "cost_per_1m_in": 0.13, + "cost_per_1m_out": 0.52, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 40960, - "default_max_tokens": 8192, + "context_window": 131072, + "default_max_tokens": 4096, "can_reason": true, "reasoning_levels": [ "low", @@ -2782,11 +2770,11 @@ { "id": "qwen/qwen3-30b-a3b-thinking-2507", "name": "Qwen: Qwen3 30B A3B Thinking 2507", - "cost_per_1m_in": 0.09, - "cost_per_1m_out": 0.3, + "cost_per_1m_in": 0.08, + "cost_per_1m_out": 0.4, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 262144, + "cost_per_1m_out_cached": 0.08, + "context_window": 131072, "default_max_tokens": 65536, "can_reason": true, "reasoning_levels": [ @@ -2837,11 +2825,11 @@ "id": "qwen/qwen3-coder-30b-a3b-instruct", "name": "Qwen: Qwen3 Coder 30B A3B Instruct", "cost_per_1m_in": 0.07, - "cost_per_1m_out": 0.28, + "cost_per_1m_out": 0.27, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 262144, - "default_max_tokens": 131072, + "context_window": 160000, + "default_max_tokens": 16384, "can_reason": false, "supports_attachments": false }, @@ -2938,12 +2926,12 @@ { "id": "qwen/qwen3-next-80b-a3b-instruct", "name": "Qwen: Qwen3 Next 80B A3B Instruct", - "cost_per_1m_in": 0.1, + "cost_per_1m_in": 0.09, "cost_per_1m_out": 1.1, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.07, + "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 8192, "can_reason": false, "supports_attachments": false }, @@ -3010,12 +2998,12 @@ { "id": "qwen/qwen3-vl-30b-a3b-instruct", "name": "Qwen: Qwen3 VL 30B A3B Instruct", - "cost_per_1m_in": 0.15, - "cost_per_1m_out": 0.6, + "cost_per_1m_in": 0.29, + "cost_per_1m_out": 1, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 8192, + "default_max_tokens": 131072, "can_reason": false, "supports_attachments": true }, @@ -3085,9 +3073,9 @@ "cost_per_1m_in": 0.6, "cost_per_1m_out": 3.6, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.3, "context_window": 262144, - "default_max_tokens": 32768, + "default_max_tokens": 131072, "can_reason": true, "reasoning_levels": [ "low", @@ -3136,10 +3124,10 @@ { "id": "qwen/qwen3.5-122b-a10b", "name": "Qwen: Qwen3.5-122B-A10B", - "cost_per_1m_in": 0.4, - "cost_per_1m_out": 3.2, + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 2.4, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.3, "context_window": 262144, "default_max_tokens": 32768, "can_reason": true, @@ -3172,12 +3160,12 @@ { "id": "qwen/qwen3.5-35b-a3b", "name": "Qwen: Qwen3.5-35B-A3B", - "cost_per_1m_in": 0.225, - "cost_per_1m_out": 1.8, + "cost_per_1m_in": 0.2, + "cost_per_1m_out": 1, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.225, + "cost_per_1m_out_cached": 0.05, "context_window": 262144, - "default_max_tokens": 32768, + "default_max_tokens": 131072, "can_reason": true, "reasoning_levels": [ "low", @@ -3226,12 +3214,12 @@ { "id": "qwen/qwen3.6-27b", "name": "Qwen: Qwen3.6 27B", - "cost_per_1m_in": 0.32, - "cost_per_1m_out": 3.2, + "cost_per_1m_in": 0.6, + "cost_per_1m_out": 3.6, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 40960, + "default_max_tokens": 32768, "can_reason": true, "reasoning_levels": [ "low", @@ -3531,7 +3519,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.11, "context_window": 131072, - "default_max_tokens": 48000, + "default_max_tokens": 49152, "can_reason": true, "reasoning_levels": [ "low", @@ -3766,10 +3754,10 @@ { "id": "x-ai/grok-3", "name": "xAI: Grok 3", - "cost_per_1m_in": 5, - "cost_per_1m_out": 25, + "cost_per_1m_in": 3, + "cost_per_1m_out": 15, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 1.25, + "cost_per_1m_out_cached": 0.75, "context_window": 131072, "default_max_tokens": 13107, "can_reason": false, @@ -3778,10 +3766,10 @@ { "id": "x-ai/grok-3-beta", "name": "xAI: Grok 3 Beta", - "cost_per_1m_in": 5, - "cost_per_1m_out": 25, + "cost_per_1m_in": 3, + "cost_per_1m_out": 15, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 1.25, + "cost_per_1m_out_cached": 0.75, "context_window": 131072, "default_max_tokens": 13107, "can_reason": false,