From da51731d2afedc0b9599144b486c8afec02e2fae Mon Sep 17 00:00:00 2001 From: Charm <124303983+charmcli@users.noreply.github.com> Date: Tue, 5 May 2026 03:12:17 +0000 Subject: [PATCH] chore: auto-update generated files --- internal/providers/configs/openrouter.json | 208 +++++++++++---------- internal/providers/configs/venice.json | 14 +- 2 files changed, 114 insertions(+), 108 deletions(-) diff --git a/internal/providers/configs/openrouter.json b/internal/providers/configs/openrouter.json index 8a8365eeafb3d749f0d03bae258fdc6746b6210f..19366d08b2e3bba7f391c60921ae81fe9152c808 100644 --- a/internal/providers/configs/openrouter.json +++ b/internal/providers/configs/openrouter.json @@ -526,36 +526,36 @@ { "id": "deepseek/deepseek-chat", "name": "DeepSeek: DeepSeek V3", - "cost_per_1m_in": 0.32, - "cost_per_1m_out": 0.89, + "cost_per_1m_in": 0.4, + "cost_per_1m_out": 1.3, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 163840, - "default_max_tokens": 8192, + "context_window": 64000, + "default_max_tokens": 8000, "can_reason": false, "supports_attachments": false }, { "id": "deepseek/deepseek-chat-v3-0324", "name": "DeepSeek: DeepSeek V3 0324", - "cost_per_1m_in": 0.216, - "cost_per_1m_out": 0.88, + "cost_per_1m_in": 0.27, + "cost_per_1m_out": 1.12, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.18, - "context_window": 131072, - "default_max_tokens": 8192, + "cost_per_1m_out_cached": 0.135, + "context_window": 163840, + "default_max_tokens": 81920, "can_reason": false, "supports_attachments": false }, { "id": "deepseek/deepseek-chat-v3.1", "name": "DeepSeek: DeepSeek V3.1", - "cost_per_1m_in": 0.56, - "cost_per_1m_out": 1.68, + "cost_per_1m_in": 0.27, + "cost_per_1m_out": 1, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.28, + "cost_per_1m_out_cached": 0.135, "context_window": 163840, - "default_max_tokens": 16384, + "default_max_tokens": 32768, "can_reason": true, "reasoning_levels": [ "low", @@ -569,11 +569,11 @@ "id": "deepseek/deepseek-v3.1-terminus", "name": "DeepSeek: DeepSeek V3.1 Terminus", "cost_per_1m_in": 0.27, - "cost_per_1m_out": 0.95, + "cost_per_1m_out": 1, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.13, + "cost_per_1m_out_cached": 0, "context_window": 163840, - "default_max_tokens": 16384, + "default_max_tokens": 81920, "can_reason": true, "reasoning_levels": [ "low", @@ -586,12 +586,12 @@ { "id": "deepseek/deepseek-v3.2", "name": "DeepSeek: DeepSeek V3.2", - "cost_per_1m_in": 0.269, - "cost_per_1m_out": 0.4, + "cost_per_1m_in": 0.26, + "cost_per_1m_out": 0.38, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.1345, + "cost_per_1m_out_cached": 0.13, "context_window": 163840, - "default_max_tokens": 32768, + "default_max_tokens": 8192, "can_reason": true, "reasoning_levels": [ "low", @@ -640,12 +640,12 @@ { "id": "deepseek/deepseek-v4-pro", "name": "DeepSeek: DeepSeek V4 Pro", - "cost_per_1m_in": 1.74, - "cost_per_1m_out": 3.48, + "cost_per_1m_in": 0.435, + "cost_per_1m_out": 0.87, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.145, + "cost_per_1m_out_cached": 0.00363, "context_window": 1048576, - "default_max_tokens": 196608, + "default_max_tokens": 192000, "can_reason": true, "reasoning_levels": [ "low", @@ -676,12 +676,12 @@ { "id": "deepseek/deepseek-r1-0528", "name": "DeepSeek: R1 0528", - "cost_per_1m_in": 0.55, - "cost_per_1m_out": 2.15, + "cost_per_1m_in": 0.5, + "cost_per_1m_out": 2.18, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.35, - "context_window": 131072, - "default_max_tokens": 65536, + "cost_per_1m_out_cached": 0, + "context_window": 163840, + "default_max_tokens": 81920, "can_reason": true, "reasoning_levels": [ "low", @@ -1097,11 +1097,11 @@ "id": "minimax/minimax-m2", "name": "MiniMax: MiniMax M2", "cost_per_1m_in": 0.255, - "cost_per_1m_out": 1, + "cost_per_1m_out": 1.02, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.03, - "context_window": 196608, - "default_max_tokens": 98304, + "cost_per_1m_out_cached": 0, + "context_window": 204800, + "default_max_tokens": 65536, "can_reason": true, "reasoning_levels": [ "low", @@ -1114,12 +1114,12 @@ { "id": "minimax/minimax-m2.1", "name": "MiniMax: MiniMax M2.1", - "cost_per_1m_in": 0.29, - "cost_per_1m_out": 0.95, + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 1.2, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.03, - "context_window": 196608, - "default_max_tokens": 98304, + "context_window": 204800, + "default_max_tokens": 65536, "can_reason": true, "reasoning_levels": [ "low", @@ -1168,8 +1168,8 @@ { "id": "minimax/minimax-m2.7", "name": "MiniMax: MiniMax M2.7", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 2.4, + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 1.2, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.06, "context_window": 204800, @@ -1282,12 +1282,12 @@ { "id": "mistralai/ministral-3b-2512", "name": "Mistral: Ministral 3 3B 2512", - "cost_per_1m_in": 0.1, - "cost_per_1m_out": 0.1, + "cost_per_1m_in": 0.15, + "cost_per_1m_out": 0.15, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.01, + "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 13107, + "default_max_tokens": 65536, "can_reason": false, "supports_attachments": true }, @@ -1339,6 +1339,24 @@ "can_reason": false, "supports_attachments": true }, + { + "id": "mistralai/mistral-medium-3-5", + "name": "Mistral: Mistral Medium 3.5", + "cost_per_1m_in": 1.5, + "cost_per_1m_out": 7.5, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 262144, + "default_max_tokens": 26214, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": true + }, { "id": "mistralai/mistral-nemo", "name": "Mistral: Mistral Nemo", @@ -1509,7 +1527,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.16, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 26214, "can_reason": true, "reasoning_levels": [ "low", @@ -1981,7 +1999,7 @@ "cost_per_1m_in": 1.25, "cost_per_1m_out": 10, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.13, + "cost_per_1m_out_cached": 0.125, "context_window": 400000, "default_max_tokens": 64000, "can_reason": true, @@ -2266,12 +2284,12 @@ { "id": "openai/gpt-oss-120b", "name": "OpenAI: gpt-oss-120b", - "cost_per_1m_in": 0.35, - "cost_per_1m_out": 0.75, + "cost_per_1m_in": 0.05, + "cost_per_1m_out": 0.25, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.35, + "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 20480, + "default_max_tokens": 16384, "can_reason": true, "reasoning_levels": [ "low", @@ -2686,12 +2704,12 @@ { "id": "qwen/qwen3-14b", "name": "Qwen: Qwen3 14B", - "cost_per_1m_in": 0.2275, - "cost_per_1m_out": 0.91, + "cost_per_1m_in": 0.12, + "cost_per_1m_out": 0.24, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 131072, - "default_max_tokens": 4096, + "context_window": 40960, + "default_max_tokens": 8192, "can_reason": true, "reasoning_levels": [ "low", @@ -2836,12 +2854,12 @@ { "id": "qwen/qwen3-coder-30b-a3b-instruct", "name": "Qwen: Qwen3 Coder 30B A3B Instruct", - "cost_per_1m_in": 0.2925, - "cost_per_1m_out": 1.4625, + "cost_per_1m_in": 0.07, + "cost_per_1m_out": 0.28, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 32768, + "default_max_tokens": 131072, "can_reason": false, "supports_attachments": false }, @@ -2884,10 +2902,10 @@ { "id": "qwen/qwen3-coder-next", "name": "Qwen: Qwen3 Coder Next", - "cost_per_1m_in": 0.14, + "cost_per_1m_in": 0.12, "cost_per_1m_out": 0.8, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.09, + "cost_per_1m_out_cached": 0.07, "context_window": 262144, "default_max_tokens": 131072, "can_reason": false, @@ -2938,10 +2956,10 @@ { "id": "qwen/qwen3-next-80b-a3b-instruct", "name": "Qwen: Qwen3 Next 80B A3B Instruct", - "cost_per_1m_in": 0.15, - "cost_per_1m_out": 1.2, + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 1.1, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.07, "context_window": 262144, "default_max_tokens": 131072, "can_reason": false, @@ -2980,11 +2998,11 @@ { "id": "qwen/qwen3-vl-235b-a22b-instruct", "name": "Qwen: Qwen3 VL 235B A22B Instruct", - "cost_per_1m_in": 0.2, - "cost_per_1m_out": 0.88, + "cost_per_1m_in": 0.25, + "cost_per_1m_out": 1.5, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.11, - "context_window": 262144, + "cost_per_1m_out_cached": 0, + "context_window": 256000, "default_max_tokens": 8192, "can_reason": false, "supports_attachments": true @@ -2992,8 +3010,8 @@ { "id": "qwen/qwen3-vl-235b-a22b-thinking", "name": "Qwen: Qwen3 VL 235B A22B Thinking", - "cost_per_1m_in": 0.26, - "cost_per_1m_out": 2.6, + "cost_per_1m_in": 0.98, + "cost_per_1m_out": 3.95, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, @@ -3022,8 +3040,8 @@ { "id": "qwen/qwen3-vl-30b-a3b-thinking", "name": "Qwen: Qwen3 VL 30B A3B Thinking", - "cost_per_1m_in": 0.2, - "cost_per_1m_out": 1, + "cost_per_1m_in": 0.13, + "cost_per_1m_out": 1.56, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, @@ -3052,12 +3070,12 @@ { "id": "qwen/qwen3-vl-8b-instruct", "name": "Qwen: Qwen3 VL 8B Instruct", - "cost_per_1m_in": 0.117, - "cost_per_1m_out": 0.455, + "cost_per_1m_in": 0.25, + "cost_per_1m_out": 0.75, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 131072, - "default_max_tokens": 16384, + "cost_per_1m_out_cached": 0.12, + "context_window": 262144, + "default_max_tokens": 131072, "can_reason": false, "supports_attachments": true }, @@ -3082,12 +3100,12 @@ { "id": "qwen/qwen3.5-397b-a17b", "name": "Qwen: Qwen3.5 397B A17B", - "cost_per_1m_in": 0.5, - "cost_per_1m_out": 3.6, + "cost_per_1m_in": 0.39, + "cost_per_1m_out": 2.34, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.3, + "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 32768, "can_reason": true, "reasoning_levels": [ "low", @@ -3136,10 +3154,10 @@ { "id": "qwen/qwen3.5-122b-a10b", "name": "Qwen: Qwen3.5-122B-A10B", - "cost_per_1m_in": 0.3, - "cost_per_1m_out": 2.4, + "cost_per_1m_in": 0.26, + "cost_per_1m_out": 2.08, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.3, + "cost_per_1m_out_cached": 0, "context_window": 262144, "default_max_tokens": 32768, "can_reason": true, @@ -3172,10 +3190,10 @@ { "id": "qwen/qwen3.5-35b-a3b", "name": "Qwen: Qwen3.5-35B-A3B", - "cost_per_1m_in": 0.23, - "cost_per_1m_out": 1.8, + "cost_per_1m_in": 0.15, + "cost_per_1m_out": 1, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.05, "context_window": 262144, "default_max_tokens": 131072, "can_reason": true, @@ -3652,12 +3670,12 @@ { "id": "z-ai/glm-4.7", "name": "Z.ai: GLM 4.7", - "cost_per_1m_in": 0.45, - "cost_per_1m_out": 2.2, + "cost_per_1m_in": 0.52, + "cost_per_1m_out": 1.85, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.11, - "context_window": 204800, - "default_max_tokens": 102400, + "cost_per_1m_out_cached": 0.12, + "context_window": 202752, + "default_max_tokens": 101376, "can_reason": true, "reasoning_levels": [ "low", @@ -3784,10 +3802,10 @@ { "id": "x-ai/grok-3", "name": "xAI: Grok 3", - "cost_per_1m_in": 5, - "cost_per_1m_out": 25, + "cost_per_1m_in": 3, + "cost_per_1m_out": 15, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 1.25, + "cost_per_1m_out_cached": 0.75, "context_window": 131072, "default_max_tokens": 13107, "can_reason": false, @@ -3796,10 +3814,10 @@ { "id": "x-ai/grok-3-beta", "name": "xAI: Grok 3 Beta", - "cost_per_1m_in": 5, - "cost_per_1m_out": 25, + "cost_per_1m_in": 3, + "cost_per_1m_out": 15, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 1.25, + "cost_per_1m_out_cached": 0.75, "context_window": 131072, "default_max_tokens": 13107, "can_reason": false, diff --git a/internal/providers/configs/venice.json b/internal/providers/configs/venice.json index e39be1d24390ffeb37d21eeb4b01ec4f227a4fde..f2fc61e7d81b621c923363017123e0bc48c9adc2 100644 --- a/internal/providers/configs/venice.json +++ b/internal/providers/configs/venice.json @@ -760,7 +760,7 @@ "cost_per_1m_out_cached": 0, "context_window": 128000, "default_max_tokens": 16384, - "can_reason": false, + "can_reason": true, "supports_attachments": false }, { @@ -799,18 +799,6 @@ "can_reason": false, "supports_attachments": false }, - { - "id": "qwen3-coder-480b-a35b-instruct", - "name": "Qwen 3 Coder 480b", - "cost_per_1m_in": 0.75, - "cost_per_1m_out": 3, - "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 256000, - "default_max_tokens": 65536, - "can_reason": false, - "supports_attachments": false - }, { "id": "qwen3-next-80b", "name": "Qwen 3 Next 80b",