From cfb625ce7583f8432c5ddbb22d8e49ce46b50dc4 Mon Sep 17 00:00:00 2001 From: Charm <124303983+charmcli@users.noreply.github.com> Date: Sun, 10 May 2026 03:13:42 +0000 Subject: [PATCH] chore: auto-update generated files --- internal/providers/configs/neuralwatt.json | 8 +- internal/providers/configs/openrouter.json | 128 ++++++++++----------- 2 files changed, 62 insertions(+), 74 deletions(-) diff --git a/internal/providers/configs/neuralwatt.json b/internal/providers/configs/neuralwatt.json index 6dddd787e14ec527e4bbb6b196c5d804dc539d41..5dddac9bc9655753255d89280df0040a19a05c22 100644 --- a/internal/providers/configs/neuralwatt.json +++ b/internal/providers/configs/neuralwatt.json @@ -146,8 +146,8 @@ "cost_per_1m_out": 0.1, "cost_per_1m_in_cached": 0.05, "cost_per_1m_out_cached": 0.1, - "context_window": 65536, - "default_max_tokens": 6553, + "context_window": 131072, + "default_max_tokens": 13107, "can_reason": true, "supports_attachments": true }, @@ -158,8 +158,8 @@ "cost_per_1m_out": 0.1, "cost_per_1m_in_cached": 0.05, "cost_per_1m_out_cached": 0.1, - "context_window": 65536, - "default_max_tokens": 6553, + "context_window": 131072, + "default_max_tokens": 13107, "can_reason": false, "supports_attachments": true } diff --git a/internal/providers/configs/openrouter.json b/internal/providers/configs/openrouter.json index da18747e9e41dc6964675106fb00aa4ac92969e9..df6e9a5c1a907928e351e4928281a6e5db10cb12 100644 --- a/internal/providers/configs/openrouter.json +++ b/internal/providers/configs/openrouter.json @@ -278,7 +278,7 @@ "cost_per_1m_out": 15, "cost_per_1m_in_cached": 3.75, "cost_per_1m_out_cached": 0.3, - "context_window": 1000000, + "context_window": 200000, "default_max_tokens": 32000, "can_reason": true, "reasoning_levels": [ @@ -544,12 +544,12 @@ { "id": "deepseek/deepseek-chat-v3-0324", "name": "DeepSeek: DeepSeek V3 0324", - "cost_per_1m_in": 0.25, - "cost_per_1m_out": 1, + "cost_per_1m_in": 0.27, + "cost_per_1m_out": 1.12, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.135, "context_window": 163840, - "default_max_tokens": 81920, + "default_max_tokens": 32768, "can_reason": false, "supports_attachments": false }, @@ -575,11 +575,11 @@ "id": "deepseek/deepseek-v3.1-terminus", "name": "DeepSeek: DeepSeek V3.1 Terminus", "cost_per_1m_in": 0.27, - "cost_per_1m_out": 1, + "cost_per_1m_out": 0.95, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.13, "context_window": 163840, - "default_max_tokens": 81920, + "default_max_tokens": 16384, "can_reason": true, "reasoning_levels": [ "low", @@ -958,12 +958,12 @@ { "id": "google/gemma-4-26b-a4b-it", "name": "Google: Gemma 4 26B A4B ", - "cost_per_1m_in": 0.15, - "cost_per_1m_out": 0.4, + "cost_per_1m_in": 0.07, + "cost_per_1m_out": 0.34, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.1, + "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 8192, "can_reason": true, "reasoning_levels": [ "low", @@ -1306,12 +1306,12 @@ { "id": "mistralai/ministral-3b-2512", "name": "Mistral: Ministral 3 3B 2512", - "cost_per_1m_in": 0.15, - "cost_per_1m_out": 0.15, + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.1, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.01, "context_window": 131072, - "default_max_tokens": 65536, + "default_max_tokens": 13107, "can_reason": false, "supports_attachments": true }, @@ -1486,12 +1486,12 @@ { "id": "moonshotai/kimi-k2-0905", "name": "MoonshotAI: Kimi K2 0905", - "cost_per_1m_in": 1, - "cost_per_1m_out": 3, + "cost_per_1m_in": 0.6, + "cost_per_1m_out": 2.5, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.5, + "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 8192, + "default_max_tokens": 131072, "can_reason": false, "supports_attachments": false }, @@ -1517,11 +1517,11 @@ "id": "moonshotai/kimi-k2.5", "name": "MoonshotAI: Kimi K2.5", "cost_per_1m_in": 0.6, - "cost_per_1m_out": 3, + "cost_per_1m_out": 2.8, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.1, + "cost_per_1m_out_cached": 0.2, "context_window": 262144, - "default_max_tokens": 26214, + "default_max_tokens": 131072, "can_reason": true, "reasoning_levels": [ "low", @@ -2077,9 +2077,9 @@ "cost_per_1m_in": 0.25, "cost_per_1m_out": 2, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.03, + "cost_per_1m_out_cached": 0.025, "context_window": 400000, - "default_max_tokens": 64000, + "default_max_tokens": 50000, "can_reason": true, "reasoning_levels": [ "low", @@ -2296,12 +2296,12 @@ { "id": "openai/gpt-oss-120b", "name": "OpenAI: gpt-oss-120b", - "cost_per_1m_in": 0.05, - "cost_per_1m_out": 0.25, + "cost_per_1m_in": 0.039, + "cost_per_1m_out": 0.19, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 16384, + "default_max_tokens": 65536, "can_reason": true, "reasoning_levels": [ "low", @@ -2701,18 +2701,6 @@ "can_reason": false, "supports_attachments": false }, - { - "id": "qwen/qwen-2.5-7b-instruct", - "name": "Qwen: Qwen2.5 7B Instruct", - "cost_per_1m_in": 0.04, - "cost_per_1m_out": 0.1, - "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.04, - "context_window": 32768, - "default_max_tokens": 4096, - "can_reason": false, - "supports_attachments": false - }, { "id": "qwen/qwen3-14b", "name": "Qwen: Qwen3 14B", @@ -2752,10 +2740,10 @@ { "id": "qwen/qwen3-235b-a22b-2507", "name": "Qwen: Qwen3 235B A22B Instruct 2507", - "cost_per_1m_in": 0.1, - "cost_per_1m_out": 0.1, + "cost_per_1m_in": 0.2, + "cost_per_1m_out": 0.8, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.1, + "cost_per_1m_out_cached": 0, "context_window": 262144, "default_max_tokens": 131072, "can_reason": false, @@ -2812,11 +2800,11 @@ { "id": "qwen/qwen3-30b-a3b-thinking-2507", "name": "Qwen: Qwen3 30B A3B Thinking 2507", - "cost_per_1m_in": 0.09, - "cost_per_1m_out": 0.3, + "cost_per_1m_in": 0.08, + "cost_per_1m_out": 0.4, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 262144, + "cost_per_1m_out_cached": 0.08, + "context_window": 131072, "default_max_tokens": 65536, "can_reason": true, "reasoning_levels": [ @@ -2878,8 +2866,8 @@ { "id": "qwen/qwen3-coder", "name": "Qwen: Qwen3 Coder 480B A35B", - "cost_per_1m_in": 0.22, - "cost_per_1m_out": 1.8, + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 1.3, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, @@ -2914,10 +2902,10 @@ { "id": "qwen/qwen3-coder-next", "name": "Qwen: Qwen3 Coder Next", - "cost_per_1m_in": 0.11, - "cost_per_1m_out": 0.8, + "cost_per_1m_in": 0.18, + "cost_per_1m_out": 1.35, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.07, + "cost_per_1m_out_cached": 0.18, "context_window": 262144, "default_max_tokens": 131072, "can_reason": false, @@ -3112,12 +3100,12 @@ { "id": "qwen/qwen3.5-397b-a17b", "name": "Qwen: Qwen3.5 397B A17B", - "cost_per_1m_in": 0.6, + "cost_per_1m_in": 0.5, "cost_per_1m_out": 3.6, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.3, "context_window": 262144, - "default_max_tokens": 26214, + "default_max_tokens": 131072, "can_reason": true, "reasoning_levels": [ "low", @@ -3166,10 +3154,10 @@ { "id": "qwen/qwen3.5-122b-a10b", "name": "Qwen: Qwen3.5-122B-A10B", - "cost_per_1m_in": 0.3, - "cost_per_1m_out": 2.4, + "cost_per_1m_in": 0.26, + "cost_per_1m_out": 2.08, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.3, + "cost_per_1m_out_cached": 0, "context_window": 262144, "default_max_tokens": 32768, "can_reason": true, @@ -3202,12 +3190,12 @@ { "id": "qwen/qwen3.5-35b-a3b", "name": "Qwen: Qwen3.5-35B-A3B", - "cost_per_1m_in": 0.19, + "cost_per_1m_in": 0.15, "cost_per_1m_out": 1, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.05, "context_window": 262144, - "default_max_tokens": 26214, + "default_max_tokens": 131072, "can_reason": true, "reasoning_levels": [ "low", @@ -3274,10 +3262,10 @@ { "id": "qwen/qwen3.6-35b-a3b", "name": "Qwen: Qwen3.6 35B A3B", - "cost_per_1m_in": 0.23, - "cost_per_1m_out": 1.8, + "cost_per_1m_in": 0.15, + "cost_per_1m_out": 1, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.05, "context_window": 262144, "default_max_tokens": 131072, "can_reason": true, @@ -3561,7 +3549,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.11, "context_window": 131072, - "default_max_tokens": 49152, + "default_max_tokens": 48000, "can_reason": true, "reasoning_levels": [ "low", @@ -3664,12 +3652,12 @@ { "id": "z-ai/glm-4.7", "name": "Z.ai: GLM 4.7", - "cost_per_1m_in": 0.52, - "cost_per_1m_out": 1.85, + "cost_per_1m_in": 0.45, + "cost_per_1m_out": 2.2, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.12, - "context_window": 202752, - "default_max_tokens": 101376, + "cost_per_1m_out_cached": 0.11, + "context_window": 204800, + "default_max_tokens": 102400, "can_reason": true, "reasoning_levels": [ "low",