From 1b12784eab14fd4719a936a9e3a79876d1914041 Mon Sep 17 00:00:00 2001 From: Charm <124303983+charmcli@users.noreply.github.com> Date: Sat, 11 Apr 2026 03:08:20 +0000 Subject: [PATCH] chore: auto-update generated files --- internal/providers/configs/aihubmix.json | 22 ++- internal/providers/configs/chutes.json | 30 --- internal/providers/configs/cortecs.json | 120 ++++++------ internal/providers/configs/huggingface.json | 24 +-- internal/providers/configs/openrouter.json | 200 +++++++++----------- 5 files changed, 183 insertions(+), 213 deletions(-) diff --git a/internal/providers/configs/aihubmix.json b/internal/providers/configs/aihubmix.json index 1dc056af0dba1b56fa6fc8fddb9be849a143ebc2..003ce25ff6e1a991cab76ba8f3f933bb772b1f24 100644 --- a/internal/providers/configs/aihubmix.json +++ b/internal/providers/configs/aihubmix.json @@ -2131,7 +2131,7 @@ "cost_per_1m_in": 2, "cost_per_1m_out": 6, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.5, + "cost_per_1m_out_cached": 0.2, "context_window": 2000000, "default_max_tokens": 200000, "can_reason": true, @@ -2149,7 +2149,7 @@ "cost_per_1m_in": 2, "cost_per_1m_out": 6, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.5, + "cost_per_1m_out_cached": 0.2, "context_window": 2000000, "default_max_tokens": 200000, "can_reason": true, @@ -2227,6 +2227,24 @@ "default_reasoning_effort": "medium", "supports_attachments": true }, + { + "id": "grok-4.20-multi-agent-0309", + "name": "Grok 4.20 Multi Agent 0309", + "cost_per_1m_in": 2, + "cost_per_1m_out": 6, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0.2, + "context_window": 2000000, + "default_max_tokens": 200000, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": true + }, { "id": "grok-4.20-multi-agent-beta-0309", "name": "Grok 4.20 Multi Agent Beta 0309", diff --git a/internal/providers/configs/chutes.json b/internal/providers/configs/chutes.json index 14ec43c2ed74ada3afb02fdd0400f40f108b0d63..5b8253d2636fc5ff024339e6884c4976580d76f9 100644 --- a/internal/providers/configs/chutes.json +++ b/internal/providers/configs/chutes.json @@ -361,18 +361,6 @@ "can_reason": false, "supports_attachments": true }, - { - "id": "chutesai/Mistral-Small-3.1-24B-Instruct-2503-TEE", - "name": "Mistral-Small-3.1-24B-Instruct-2503-TEE", - "cost_per_1m_in": 0.03, - "cost_per_1m_out": 0.11, - "cost_per_1m_in_cached": 0.015, - "cost_per_1m_out_cached": 0, - "context_window": 131072, - "default_max_tokens": 131072, - "can_reason": false, - "supports_attachments": true - }, { "id": "chutesai/Mistral-Small-3.2-24B-Instruct-2506", "name": "Mistral-Small-3.2-24B-Instruct-2506", @@ -534,24 +522,6 @@ ], "default_reasoning_effort": "medium", "supports_attachments": false - }, - { - "id": "openai/gpt-oss-20b-TEE", - "name": "gpt-oss-20b-TEE", - "cost_per_1m_in": 0.03, - "cost_per_1m_out": 0.11, - "cost_per_1m_in_cached": 0.015, - "cost_per_1m_out_cached": 0, - "context_window": 131072, - "default_max_tokens": 131072, - "can_reason": true, - "reasoning_levels": [ - "low", - "medium", - "high" - ], - "default_reasoning_effort": "medium", - "supports_attachments": false } ] } diff --git a/internal/providers/configs/cortecs.json b/internal/providers/configs/cortecs.json index 0ff80b3acd7f3e83b32a7c2b0a059f5cd43879b2..75c4ff1cc6e91a1611db30631b3e9ba34c781fb6 100644 --- a/internal/providers/configs/cortecs.json +++ b/internal/providers/configs/cortecs.json @@ -992,10 +992,10 @@ "supports_attachments": false }, { - "id": "gpt-4.1", - "name": "GPT 4.1", - "cost_per_1m_in": 2.066, - "cost_per_1m_out": 8.266, + "id": "gpt-4.1-mini", + "name": "GPT 4.1 mini", + "cost_per_1m_in": 0.41, + "cost_per_1m_out": 1.607, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 1047576, @@ -1010,10 +1010,10 @@ "supports_attachments": true }, { - "id": "gpt-4.1-mini", - "name": "GPT 4.1 mini", - "cost_per_1m_in": 0.41, - "cost_per_1m_out": 1.607, + "id": "gpt-4.1-nano", + "name": "GPT 4.1 nano", + "cost_per_1m_in": 0.105, + "cost_per_1m_out": 0.41, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 1047576, @@ -1028,14 +1028,14 @@ "supports_attachments": true }, { - "id": "gpt-4.1-nano", - "name": "GPT 4.1 nano", - "cost_per_1m_in": 0.105, - "cost_per_1m_out": 0.41, + "id": "nova-lite-v1", + "name": "Nova Lite 1.0", + "cost_per_1m_in": 0.065, + "cost_per_1m_out": 0.259, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 1047576, - "default_max_tokens": 104757, + "context_window": 300000, + "default_max_tokens": 30000, "can_reason": true, "reasoning_levels": [ "low", @@ -1064,14 +1064,14 @@ "supports_attachments": true }, { - "id": "nova-lite-v1", - "name": "Nova Lite 1.0", - "cost_per_1m_in": 0.065, - "cost_per_1m_out": 0.259, + "id": "gpt-4.1", + "name": "GPT 4.1", + "cost_per_1m_in": 2.066, + "cost_per_1m_out": 8.266, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 300000, - "default_max_tokens": 30000, + "context_window": 1047576, + "default_max_tokens": 104757, "can_reason": true, "reasoning_levels": [ "low", @@ -1082,14 +1082,14 @@ "supports_attachments": true }, { - "id": "nova-pro-v1", - "name": "Nova Pro 1.0", - "cost_per_1m_in": 0.865, - "cost_per_1m_out": 3.46, + "id": "claude-3-7-sonnet", + "name": "Claude 3.7 Sonnet", + "cost_per_1m_in": 2.817, + "cost_per_1m_out": 14.087, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 300000, - "default_max_tokens": 30000, + "context_window": 200000, + "default_max_tokens": 20000, "can_reason": true, "reasoning_levels": [ "low", @@ -1100,14 +1100,14 @@ "supports_attachments": true }, { - "id": "claude-sonnet-4", - "name": "Claude Sonnet 4", - "cost_per_1m_in": 2.817, - "cost_per_1m_out": 14.087, + "id": "nova-pro-v1", + "name": "Nova Pro 1.0", + "cost_per_1m_in": 0.865, + "cost_per_1m_out": 3.46, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 200000, - "default_max_tokens": 20000, + "context_window": 300000, + "default_max_tokens": 30000, "can_reason": true, "reasoning_levels": [ "low", @@ -1118,8 +1118,8 @@ "supports_attachments": true }, { - "id": "claude-3-7-sonnet", - "name": "Claude 3.7 Sonnet", + "id": "claude-sonnet-4", + "name": "Claude Sonnet 4", "cost_per_1m_in": 2.817, "cost_per_1m_out": 14.087, "cost_per_1m_in_cached": 0, @@ -1370,10 +1370,10 @@ "supports_attachments": true }, { - "id": "gpt-5-mini", - "name": "GPT 5 mini", - "cost_per_1m_in": 0.263, - "cost_per_1m_out": 2.066, + "id": "gpt-5-nano", + "name": "GPT 5 nano", + "cost_per_1m_in": 0.057, + "cost_per_1m_out": 0.414, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 400000, @@ -1388,10 +1388,10 @@ "supports_attachments": true }, { - "id": "gpt-5-nano", - "name": "GPT 5 nano", - "cost_per_1m_in": 0.057, - "cost_per_1m_out": 0.414, + "id": "gpt-5-mini", + "name": "GPT 5 mini", + "cost_per_1m_in": 0.263, + "cost_per_1m_out": 2.066, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 400000, @@ -1447,6 +1447,24 @@ "can_reason": false, "supports_attachments": false }, + { + "id": "mistral-medium-2508", + "name": "Mistral Medium 2508", + "cost_per_1m_in": 0.42, + "cost_per_1m_out": 2.1, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 128000, + "default_max_tokens": 12800, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": true + }, { "id": "devstral-medium-2507", "name": "Devstral Medium 2507", @@ -1471,24 +1489,6 @@ "can_reason": false, "supports_attachments": false }, - { - "id": "mistral-medium-2508", - "name": "Mistral Medium 2508", - "cost_per_1m_in": 0.42, - "cost_per_1m_out": 2.1, - "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 128000, - "default_max_tokens": 12800, - "can_reason": true, - "reasoning_levels": [ - "low", - "medium", - "high" - ], - "default_reasoning_effort": "medium", - "supports_attachments": true - }, { "id": "llama-3.1-405b-instruct", "name": "Llama 3.1 405B Instruct", diff --git a/internal/providers/configs/huggingface.json b/internal/providers/configs/huggingface.json index 90e4267c12e816346eecbb2ed35971c3dbc0d74f..8261c3998ee8cfc2d3678d660bb01af338a5fda4 100644 --- a/internal/providers/configs/huggingface.json +++ b/internal/providers/configs/huggingface.json @@ -7,6 +7,18 @@ "default_large_model_id": "moonshotai/Kimi-K2.5:fireworks-ai", "default_small_model_id": "openai/gpt-oss-20b:groq", "models": [ + { + "id": "MiniMaxAI/MiniMax-M2.1:fireworks-ai", + "name": "MiniMaxAI/MiniMax-M2.1 (fireworks-ai)", + "cost_per_1m_in": 0, + "cost_per_1m_out": 0, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 204800, + "default_max_tokens": 8192, + "can_reason": false, + "supports_attachments": false + }, { "id": "MiniMaxAI/MiniMax-M2.5:fireworks-ai", "name": "MiniMaxAI/MiniMax-M2.5 (fireworks-ai)", @@ -163,18 +175,6 @@ "can_reason": false, "supports_attachments": false }, - { - "id": "openai/gpt-oss-120b:cerebras", - "name": "openai/gpt-oss-120b (cerebras)", - "cost_per_1m_in": 0.25, - "cost_per_1m_out": 0.69, - "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 131072, - "default_max_tokens": 8192, - "can_reason": false, - "supports_attachments": false - }, { "id": "openai/gpt-oss-120b:fireworks-ai", "name": "openai/gpt-oss-120b (fireworks-ai)", diff --git a/internal/providers/configs/openrouter.json b/internal/providers/configs/openrouter.json index 8241206515bd14f9405d9f2d6687c934ffa49f14..3d3b514131c0fc94e86d41f24cf26f1b52c97201 100644 --- a/internal/providers/configs/openrouter.json +++ b/internal/providers/configs/openrouter.json @@ -129,7 +129,7 @@ "cost_per_1m_in_cached": 3.75, "cost_per_1m_out_cached": 0.3, "context_window": 200000, - "default_max_tokens": 64000, + "default_max_tokens": 32000, "can_reason": true, "reasoning_levels": [ "low", @@ -532,10 +532,10 @@ { "id": "deepseek/deepseek-chat-v3.1", "name": "DeepSeek: DeepSeek V3.1", - "cost_per_1m_in": 0.56, - "cost_per_1m_out": 1.68, + "cost_per_1m_in": 0.6, + "cost_per_1m_out": 1.7, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.28, + "cost_per_1m_out_cached": 0, "context_window": 163840, "default_max_tokens": 16384, "can_reason": true, @@ -550,12 +550,12 @@ { "id": "deepseek/deepseek-v3.1-terminus", "name": "DeepSeek: DeepSeek V3.1 Terminus", - "cost_per_1m_in": 0.21, - "cost_per_1m_out": 0.79, + "cost_per_1m_in": 0.27, + "cost_per_1m_out": 1, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.13, + "cost_per_1m_out_cached": 0.135, "context_window": 163840, - "default_max_tokens": 16384, + "default_max_tokens": 32768, "can_reason": true, "reasoning_levels": [ "low", @@ -999,7 +999,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 8192, + "default_max_tokens": 13107, "can_reason": false, "supports_attachments": false }, @@ -1075,9 +1075,9 @@ "cost_per_1m_in": 0.3, "cost_per_1m_out": 1.2, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 196608, - "default_max_tokens": 19660, + "cost_per_1m_out_cached": 0.03, + "context_window": 204800, + "default_max_tokens": 65536, "can_reason": true, "reasoning_levels": [ "low", @@ -1090,12 +1090,12 @@ { "id": "minimax/minimax-m2.5", "name": "MiniMax: MiniMax M2.5", - "cost_per_1m_in": 0.2, - "cost_per_1m_out": 1.18, + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 1.2, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 196608, - "default_max_tokens": 98304, + "cost_per_1m_out_cached": 0.03, + "context_window": 204800, + "default_max_tokens": 65536, "can_reason": true, "reasoning_levels": [ "low", @@ -1252,12 +1252,12 @@ { "id": "mistralai/ministral-8b-2512", "name": "Mistral: Ministral 3 8B 2512", - "cost_per_1m_in": 0.3, - "cost_per_1m_out": 0.3, + "cost_per_1m_in": 0.15, + "cost_per_1m_out": 0.15, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.015, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 26214, "can_reason": false, "supports_attachments": true }, @@ -1426,12 +1426,12 @@ { "id": "moonshotai/kimi-k2-0905", "name": "MoonshotAI: Kimi K2 0905", - "cost_per_1m_in": 1, - "cost_per_1m_out": 3, + "cost_per_1m_in": 0.6, + "cost_per_1m_out": 2.5, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.5, + "cost_per_1m_out_cached": 0.6, "context_window": 262144, - "default_max_tokens": 8192, + "default_max_tokens": 16384, "can_reason": false, "supports_attachments": false }, @@ -1456,10 +1456,10 @@ { "id": "moonshotai/kimi-k2.5", "name": "MoonshotAI: Kimi K2.5", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 2.8, + "cost_per_1m_in": 0.44, + "cost_per_1m_out": 2.2, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.22, "context_window": 262144, "default_max_tokens": 131072, "can_reason": true, @@ -1741,7 +1741,7 @@ "cost_per_1m_in": 2.5, "cost_per_1m_out": 10, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 1.25, "context_window": 128000, "default_max_tokens": 8192, "can_reason": false, @@ -1963,7 +1963,7 @@ "cost_per_1m_in": 1.25, "cost_per_1m_out": 10, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.13, + "cost_per_1m_out_cached": 0.125, "context_window": 400000, "default_max_tokens": 64000, "can_reason": true, @@ -2212,12 +2212,12 @@ { "id": "openai/gpt-oss-120b", "name": "OpenAI: gpt-oss-120b", - "cost_per_1m_in": 0.09, - "cost_per_1m_out": 0.36, + "cost_per_1m_in": 0.039, + "cost_per_1m_out": 0.19, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.045, + "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 32768, + "default_max_tokens": 13107, "can_reason": true, "reasoning_levels": [ "low", @@ -2248,12 +2248,12 @@ { "id": "openai/gpt-oss-20b", "name": "OpenAI: gpt-oss-20b", - "cost_per_1m_in": 0.05, - "cost_per_1m_out": 0.2, + "cost_per_1m_in": 0.03, + "cost_per_1m_out": 0.14, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.05, + "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 65536, + "default_max_tokens": 13107, "can_reason": true, "reasoning_levels": [ "low", @@ -2599,24 +2599,6 @@ "can_reason": false, "supports_attachments": false }, - { - "id": "qwen/qwen3-14b", - "name": "Qwen: Qwen3 14B", - "cost_per_1m_in": 0.2275, - "cost_per_1m_out": 0.91, - "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 131072, - "default_max_tokens": 4096, - "can_reason": true, - "reasoning_levels": [ - "low", - "medium", - "high" - ], - "default_reasoning_effort": "medium", - "supports_attachments": false - }, { "id": "qwen/qwen3-235b-a22b", "name": "Qwen: Qwen3 235B A22B", @@ -2668,12 +2650,12 @@ { "id": "qwen/qwen3-30b-a3b", "name": "Qwen: Qwen3 30B A3B", - "cost_per_1m_in": 0.13, - "cost_per_1m_out": 0.52, + "cost_per_1m_in": 0.15, + "cost_per_1m_out": 0.6, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.15, "context_window": 131072, - "default_max_tokens": 4096, + "default_max_tokens": 65536, "can_reason": true, "reasoning_levels": [ "low", @@ -2689,9 +2671,9 @@ "cost_per_1m_in": 0.1, "cost_per_1m_out": 0.3, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.1, + "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 26214, "can_reason": false, "supports_attachments": false }, @@ -2752,12 +2734,12 @@ { "id": "qwen/qwen3-coder-30b-a3b-instruct", "name": "Qwen: Qwen3 Coder 30B A3B Instruct", - "cost_per_1m_in": 0.1, - "cost_per_1m_out": 0.3, + "cost_per_1m_in": 0.07, + "cost_per_1m_out": 0.28, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 26214, + "default_max_tokens": 131072, "can_reason": false, "supports_attachments": false }, @@ -2878,8 +2860,8 @@ { "id": "qwen/qwen3-next-80b-a3b-thinking", "name": "Qwen: Qwen3 Next 80B A3B Thinking", - "cost_per_1m_in": 0.15, - "cost_per_1m_out": 1.5, + "cost_per_1m_in": 0.0975, + "cost_per_1m_out": 0.78, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, @@ -2926,12 +2908,12 @@ { "id": "qwen/qwen3-vl-30b-a3b-instruct", "name": "Qwen: Qwen3 VL 30B A3B Instruct", - "cost_per_1m_in": 0.13, - "cost_per_1m_out": 0.52, + "cost_per_1m_in": 0.29, + "cost_per_1m_out": 1, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 131072, - "default_max_tokens": 16384, + "context_window": 262144, + "default_max_tokens": 131072, "can_reason": false, "supports_attachments": true }, @@ -3034,8 +3016,8 @@ { "id": "qwen/qwen3.5-122b-a10b", "name": "Qwen: Qwen3.5-122B-A10B", - "cost_per_1m_in": 0.4, - "cost_per_1m_out": 3.2, + "cost_per_1m_in": 0.26, + "cost_per_1m_out": 2.08, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, @@ -3052,10 +3034,10 @@ { "id": "qwen/qwen3.5-27b", "name": "Qwen: Qwen3.5-27B", - "cost_per_1m_in": 0.27, - "cost_per_1m_out": 2.16, + "cost_per_1m_in": 0.195, + "cost_per_1m_out": 1.56, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.27, + "cost_per_1m_out_cached": 0, "context_window": 262144, "default_max_tokens": 32768, "can_reason": true, @@ -3070,12 +3052,12 @@ { "id": "qwen/qwen3.5-35b-a3b", "name": "Qwen: Qwen3.5-35B-A3B", - "cost_per_1m_in": 0.225, - "cost_per_1m_out": 1.8, + "cost_per_1m_in": 0.25, + "cost_per_1m_out": 1, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.225, + "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 32768, + "default_max_tokens": 131072, "can_reason": true, "reasoning_levels": [ "low", @@ -3253,9 +3235,9 @@ "cost_per_1m_in": 0.1, "cost_per_1m_out": 0.3, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.1, + "cost_per_1m_out_cached": 0.02, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 16000, "can_reason": true, "reasoning_levels": [ "low", @@ -3319,9 +3301,9 @@ "cost_per_1m_in": 0.6, "cost_per_1m_out": 2.2, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.11, + "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 48000, + "default_max_tokens": 13107, "can_reason": true, "reasoning_levels": [ "low", @@ -3388,12 +3370,12 @@ { "id": "z-ai/glm-4.6", "name": "Z.ai: GLM 4.6", - "cost_per_1m_in": 0.55, - "cost_per_1m_out": 2.2, + "cost_per_1m_in": 0.39, + "cost_per_1m_out": 1.9, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.11, + "cost_per_1m_out_cached": 0, "context_window": 204800, - "default_max_tokens": 65536, + "default_max_tokens": 102400, "can_reason": true, "reasoning_levels": [ "low", @@ -3424,12 +3406,12 @@ { "id": "z-ai/glm-4.7", "name": "Z.ai: GLM 4.7", - "cost_per_1m_in": 0.54, - "cost_per_1m_out": 1.98, + "cost_per_1m_in": 0.45, + "cost_per_1m_out": 2.2, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.099, + "cost_per_1m_out_cached": 0.11, "context_window": 204800, - "default_max_tokens": 65536, + "default_max_tokens": 102400, "can_reason": true, "reasoning_levels": [ "low", @@ -3460,11 +3442,11 @@ { "id": "z-ai/glm-5", "name": "Z.ai: GLM 5", - "cost_per_1m_in": 1, - "cost_per_1m_out": 3.2, + "cost_per_1m_in": 0.95, + "cost_per_1m_out": 2.55, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.2, - "context_window": 202800, + "context_window": 204800, "default_max_tokens": 65536, "can_reason": true, "reasoning_levels": [ @@ -3500,8 +3482,8 @@ "cost_per_1m_out": 4.4, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.26, - "context_window": 202752, - "default_max_tokens": 101376, + "context_window": 204800, + "default_max_tokens": 65536, "can_reason": true, "reasoning_levels": [ "low", @@ -3532,10 +3514,10 @@ { "id": "x-ai/grok-3", "name": "xAI: Grok 3", - "cost_per_1m_in": 3, - "cost_per_1m_out": 15, + "cost_per_1m_in": 5, + "cost_per_1m_out": 25, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.75, + "cost_per_1m_out_cached": 1.25, "context_window": 131072, "default_max_tokens": 13107, "can_reason": false, @@ -3544,10 +3526,10 @@ { "id": "x-ai/grok-3-beta", "name": "xAI: Grok 3 Beta", - "cost_per_1m_in": 3, - "cost_per_1m_out": 15, + "cost_per_1m_in": 5, + "cost_per_1m_out": 25, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.75, + "cost_per_1m_out_cached": 1.25, "context_window": 131072, "default_max_tokens": 13107, "can_reason": false, @@ -3556,10 +3538,10 @@ { "id": "x-ai/grok-3-mini", "name": "xAI: Grok 3 Mini", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 4, + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 0.5, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.15, + "cost_per_1m_out_cached": 0.075, "context_window": 131072, "default_max_tokens": 13107, "can_reason": true, @@ -3574,10 +3556,10 @@ { "id": "x-ai/grok-3-mini-beta", "name": "xAI: Grok 3 Mini Beta", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 4, + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 0.5, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.15, + "cost_per_1m_out_cached": 0.075, "context_window": 131072, "default_max_tokens": 13107, "can_reason": true,