From 29b41f27bd066173feb3d242dae3e8b5d2af1b6e Mon Sep 17 00:00:00 2001 From: Charm <124303983+charmcli@users.noreply.github.com> Date: Mon, 13 Apr 2026 03:12:18 +0000 Subject: [PATCH] chore: auto-update generated files --- internal/providers/configs/huggingface.json | 12 ++ internal/providers/configs/openrouter.json | 184 +++++++++++--------- internal/providers/configs/synthetic.json | 18 ++ 3 files changed, 131 insertions(+), 83 deletions(-) diff --git a/internal/providers/configs/huggingface.json b/internal/providers/configs/huggingface.json index 8538eeace5cc6217656fea1f526a17b145be877f..3e131bf3f80e11fa925d128f04ad620b774ce372 100644 --- a/internal/providers/configs/huggingface.json +++ b/internal/providers/configs/huggingface.json @@ -31,6 +31,18 @@ "can_reason": false, "supports_attachments": false }, + { + "id": "MiniMaxAI/MiniMax-M2.7:fireworks-ai", + "name": "MiniMaxAI/MiniMax-M2.7 (fireworks-ai)", + "cost_per_1m_in": 0, + "cost_per_1m_out": 0, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 204800, + "default_max_tokens": 8192, + "can_reason": false, + "supports_attachments": false + }, { "id": "Qwen/Qwen3-235B-A22B-Instruct-2507:cerebras", "name": "Qwen/Qwen3-235B-A22B-Instruct-2507 (cerebras)", diff --git a/internal/providers/configs/openrouter.json b/internal/providers/configs/openrouter.json index 5727fc95ffe1477ce332e2d75a9abd70ca52233d..4ec4d7182fa0bb1fedf9ca05f0ae157a093acebd 100644 --- a/internal/providers/configs/openrouter.json +++ b/internal/providers/configs/openrouter.json @@ -532,12 +532,12 @@ { "id": "deepseek/deepseek-chat-v3.1", "name": "DeepSeek: DeepSeek V3.1", - "cost_per_1m_in": 0.27, - "cost_per_1m_out": 1, + "cost_per_1m_in": 0.21, + "cost_per_1m_out": 0.79, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.135, + "cost_per_1m_out_cached": 0.13, "context_window": 163840, - "default_max_tokens": 32768, + "default_max_tokens": 16384, "can_reason": true, "reasoning_levels": [ "low", @@ -622,12 +622,12 @@ { "id": "deepseek/deepseek-r1-0528", "name": "DeepSeek: R1 0528", - "cost_per_1m_in": 0.45, - "cost_per_1m_out": 2.15, + "cost_per_1m_in": 0.5, + "cost_per_1m_out": 2.18, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.225, + "cost_per_1m_out_cached": 0, "context_window": 163840, - "default_max_tokens": 32768, + "default_max_tokens": 81920, "can_reason": true, "reasoning_levels": [ "low", @@ -789,7 +789,7 @@ "cost_per_1m_in_cached": 0.08333, "cost_per_1m_out_cached": 0.05, "context_window": 1048576, - "default_max_tokens": 32767, + "default_max_tokens": 32768, "can_reason": true, "reasoning_levels": [ "low", @@ -897,7 +897,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 65536, "can_reason": true, "reasoning_levels": [ "low", @@ -1051,24 +1051,6 @@ "can_reason": false, "supports_attachments": true }, - { - "id": "minimax/minimax-m1", - "name": "MiniMax: MiniMax M1", - "cost_per_1m_in": 0.44, - "cost_per_1m_out": 1.76, - "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 1000000, - "default_max_tokens": 20000, - "can_reason": true, - "reasoning_levels": [ - "low", - "medium", - "high" - ], - "default_reasoning_effort": "medium", - "supports_attachments": false - }, { "id": "minimax/minimax-m2", "name": "MiniMax: MiniMax M2", @@ -1093,9 +1075,9 @@ "cost_per_1m_in": 0.3, "cost_per_1m_out": 1.2, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.03, - "context_window": 204800, - "default_max_tokens": 65536, + "cost_per_1m_out_cached": 0, + "context_window": 196608, + "default_max_tokens": 19660, "can_reason": true, "reasoning_levels": [ "low", @@ -1131,7 +1113,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 196608, - "default_max_tokens": 98304, + "default_max_tokens": 4096, "can_reason": true, "reasoning_levels": [ "low", @@ -1144,8 +1126,8 @@ { "id": "minimax/minimax-m2.7", "name": "MiniMax: MiniMax M2.7", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 2.4, + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 1.2, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.06, "context_window": 204800, @@ -1258,24 +1240,24 @@ { "id": "mistralai/ministral-3b-2512", "name": "Mistral: Ministral 3 3B 2512", - "cost_per_1m_in": 0.1, - "cost_per_1m_out": 0.1, + "cost_per_1m_in": 0.35, + "cost_per_1m_out": 0.35, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.01, + "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 13107, + "default_max_tokens": 65536, "can_reason": false, "supports_attachments": true }, { "id": "mistralai/ministral-8b-2512", "name": "Mistral: Ministral 3 8B 2512", - "cost_per_1m_in": 0.3, - "cost_per_1m_out": 0.3, + "cost_per_1m_in": 0.15, + "cost_per_1m_out": 0.15, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.015, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 26214, "can_reason": false, "supports_attachments": true }, @@ -1444,12 +1426,12 @@ { "id": "moonshotai/kimi-k2-0905", "name": "MoonshotAI: Kimi K2 0905", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 2.5, + "cost_per_1m_in": 1, + "cost_per_1m_out": 3, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.5, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 8192, "can_reason": false, "supports_attachments": false }, @@ -1555,6 +1537,24 @@ "default_reasoning_effort": "medium", "supports_attachments": false }, + { + "id": "nvidia/nemotron-3-super-120b-a12b", + "name": "NVIDIA: Nemotron 3 Super", + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.5, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0.1, + "context_window": 262144, + "default_max_tokens": 26214, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false + }, { "id": "nvidia/nemotron-3-super-120b-a12b:free", "name": "NVIDIA: Nemotron 3 Super (free)", @@ -1707,7 +1707,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.5, "context_window": 1047576, - "default_max_tokens": 16384, + "default_max_tokens": 104757, "can_reason": false, "supports_attachments": true }, @@ -1729,9 +1729,9 @@ "cost_per_1m_in": 0.1, "cost_per_1m_out": 0.4, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.03, + "cost_per_1m_out_cached": 0.025, "context_window": 1047576, - "default_max_tokens": 104757, + "default_max_tokens": 16384, "can_reason": false, "supports_attachments": true }, @@ -2067,7 +2067,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.175, "context_window": 128000, - "default_max_tokens": 16000, + "default_max_tokens": 8192, "can_reason": false, "supports_attachments": true }, @@ -2271,7 +2271,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 65536, + "default_max_tokens": 4096, "can_reason": true, "reasoning_levels": [ "low", @@ -2599,6 +2599,24 @@ "can_reason": false, "supports_attachments": false }, + { + "id": "qwen/qwen3-14b", + "name": "Qwen: Qwen3 14B", + "cost_per_1m_in": 0.12, + "cost_per_1m_out": 0.24, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 40960, + "default_max_tokens": 20480, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false + }, { "id": "qwen/qwen3-235b-a22b", "name": "Qwen: Qwen3 235B A22B", @@ -2680,11 +2698,11 @@ { "id": "qwen/qwen3-30b-a3b-thinking-2507", "name": "Qwen: Qwen3 30B A3B Thinking 2507", - "cost_per_1m_in": 0.09, - "cost_per_1m_out": 0.3, + "cost_per_1m_in": 0.08, + "cost_per_1m_out": 0.4, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 262144, + "cost_per_1m_out_cached": 0.08, + "context_window": 131072, "default_max_tokens": 65536, "can_reason": true, "reasoning_levels": [ @@ -2735,11 +2753,11 @@ "id": "qwen/qwen3-coder-30b-a3b-instruct", "name": "Qwen: Qwen3 Coder 30B A3B Instruct", "cost_per_1m_in": 0.07, - "cost_per_1m_out": 0.28, + "cost_per_1m_out": 0.27, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 262144, - "default_max_tokens": 131072, + "context_window": 160000, + "default_max_tokens": 16384, "can_reason": false, "supports_attachments": false }, @@ -2782,12 +2800,12 @@ { "id": "qwen/qwen3-coder-next", "name": "Qwen: Qwen3 Coder Next", - "cost_per_1m_in": 0.12, - "cost_per_1m_out": 0.75, + "cost_per_1m_in": 0.15, + "cost_per_1m_out": 0.8, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.06, + "cost_per_1m_out_cached": 0.12, "context_window": 262144, - "default_max_tokens": 32768, + "default_max_tokens": 131072, "can_reason": false, "supports_attachments": false }, @@ -2860,8 +2878,8 @@ { "id": "qwen/qwen3-next-80b-a3b-thinking", "name": "Qwen: Qwen3 Next 80B A3B Thinking", - "cost_per_1m_in": 0.0975, - "cost_per_1m_out": 0.78, + "cost_per_1m_in": 0.15, + "cost_per_1m_out": 1.5, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, @@ -2878,12 +2896,12 @@ { "id": "qwen/qwen3-vl-235b-a22b-instruct", "name": "Qwen: Qwen3 VL 235B A22B Instruct", - "cost_per_1m_in": 0.3, - "cost_per_1m_out": 1.5, + "cost_per_1m_in": 0.26, + "cost_per_1m_out": 1.04, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 262144, - "default_max_tokens": 131072, + "context_window": 131072, + "default_max_tokens": 16384, "can_reason": false, "supports_attachments": true }, @@ -2950,12 +2968,12 @@ { "id": "qwen/qwen3-vl-8b-instruct", "name": "Qwen: Qwen3 VL 8B Instruct", - "cost_per_1m_in": 0.25, - "cost_per_1m_out": 0.75, + "cost_per_1m_in": 0.117, + "cost_per_1m_out": 0.455, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 262144, - "default_max_tokens": 131072, + "context_window": 131072, + "default_max_tokens": 16384, "can_reason": false, "supports_attachments": true }, @@ -2985,7 +3003,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 32768, + "default_max_tokens": 131072, "can_reason": true, "reasoning_levels": [ "low", @@ -3139,9 +3157,9 @@ "cost_per_1m_in": 0.1, "cost_per_1m_out": 0.3, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 262144, - "default_max_tokens": 32768, + "cost_per_1m_out_cached": 0.02, + "context_window": 256000, + "default_max_tokens": 128000, "can_reason": true, "reasoning_levels": [ "low", @@ -3391,9 +3409,9 @@ "cost_per_1m_in": 0.3, "cost_per_1m_out": 0.9, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.05, "context_window": 131072, - "default_max_tokens": 65536, + "default_max_tokens": 12000, "can_reason": true, "reasoning_levels": [ "low", @@ -3442,11 +3460,11 @@ { "id": "z-ai/glm-5", "name": "Z.ai: GLM 5", - "cost_per_1m_in": 0.95, - "cost_per_1m_out": 2.55, + "cost_per_1m_in": 1, + "cost_per_1m_out": 3.2, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.2, - "context_window": 204800, + "context_window": 202800, "default_max_tokens": 65536, "can_reason": true, "reasoning_levels": [ @@ -3482,8 +3500,8 @@ "cost_per_1m_out": 4.4, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.26, - "context_window": 204800, - "default_max_tokens": 65536, + "context_window": 202752, + "default_max_tokens": 101376, "can_reason": true, "reasoning_levels": [ "low", diff --git a/internal/providers/configs/synthetic.json b/internal/providers/configs/synthetic.json index b8986320f659e1f2eb4008eac208391cd1b55268..06fa8101b3391fd684f86a42072f2b264981a36a 100644 --- a/internal/providers/configs/synthetic.json +++ b/internal/providers/configs/synthetic.json @@ -91,6 +91,24 @@ "default_reasoning_effort": "medium", "supports_attachments": false }, + { + "id": "hf:zai-org/GLM-5.1", + "name": "GLM 5.1", + "cost_per_1m_in": 1, + "cost_per_1m_out": 3, + "cost_per_1m_in_cached": 1, + "cost_per_1m_out_cached": 1, + "context_window": 196608, + "default_max_tokens": 19660, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false + }, { "id": "hf:moonshotai/Kimi-K2-Instruct-0905", "name": "Kimi K2 Instruct 0905",