diff --git a/internal/providers/configs/avian.json b/internal/providers/configs/avian.json index 558aaa5f2cd21222cdf78e7d1429f66ec83dc0a4..6bfc405fde0882c2f3c837372f46679d4f96e288 100644 --- a/internal/providers/configs/avian.json +++ b/internal/providers/configs/avian.json @@ -61,6 +61,18 @@ "can_reason": false, "supports_attachments": false }, + { + "id": "moonshotai/kimi-k2.6", + "name": "Kimi K2.6", + "cost_per_1m_in": 0.95, + "cost_per_1m_out": 4, + "cost_per_1m_in_cached": 0.16, + "cost_per_1m_out_cached": 0, + "context_window": 262144, + "default_max_tokens": 262144, + "can_reason": false, + "supports_attachments": false + }, { "id": "minimax/minimax-m2.5", "name": "MiniMax M2.5", diff --git a/internal/providers/configs/opencode-zen.json b/internal/providers/configs/opencode-zen.json index 26ddcd2623228ba238afa18d700455cebd0845cf..9e9dd93d62cd61e42af9dc43b2269a5f70dd1204 100644 --- a/internal/providers/configs/opencode-zen.json +++ b/internal/providers/configs/opencode-zen.json @@ -181,42 +181,6 @@ "default_reasoning_effort": "medium", "supports_attachments": true }, - { - "id": "glm-4.6", - "name": "GLM-4.6", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 2.2, - "cost_per_1m_in_cached": 0.1, - "cost_per_1m_out_cached": 0, - "context_window": 204800, - "default_max_tokens": 131072, - "can_reason": true, - "reasoning_levels": [ - "low", - "medium", - "high" - ], - "default_reasoning_effort": "medium", - "supports_attachments": false - }, - { - "id": "glm-4.7", - "name": "GLM-4.7", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 2.2, - "cost_per_1m_in_cached": 0.1, - "cost_per_1m_out_cached": 0, - "context_window": 204800, - "default_max_tokens": 131072, - "can_reason": true, - "reasoning_levels": [ - "low", - "medium", - "high" - ], - "default_reasoning_effort": "medium", - "supports_attachments": false - }, { "id": "glm-5", "name": "GLM-5", @@ -577,36 +541,6 @@ "default_reasoning_effort": "medium", "supports_attachments": true }, - { - "id": "kimi-k2", - "name": "Kimi K2", - "cost_per_1m_in": 0.4, - "cost_per_1m_out": 2.5, - "cost_per_1m_in_cached": 0.4, - "cost_per_1m_out_cached": 0, - "context_window": 262144, - "default_max_tokens": 262144, - "can_reason": false, - "supports_attachments": false - }, - { - "id": "kimi-k2-thinking", - "name": "Kimi K2 Thinking", - "cost_per_1m_in": 0.4, - "cost_per_1m_out": 2.5, - "cost_per_1m_in_cached": 0.4, - "cost_per_1m_out_cached": 0, - "context_window": 262144, - "default_max_tokens": 262144, - "can_reason": true, - "reasoning_levels": [ - "low", - "medium", - "high" - ], - "default_reasoning_effort": "medium", - "supports_attachments": false - }, { "id": "kimi-k2.5", "name": "Kimi K2.5", @@ -625,24 +559,6 @@ "default_reasoning_effort": "medium", "supports_attachments": true }, - { - "id": "minimax-m2.1", - "name": "MiniMax M2.1", - "cost_per_1m_in": 0.3, - "cost_per_1m_out": 1.2, - "cost_per_1m_in_cached": 0.1, - "cost_per_1m_out_cached": 0, - "context_window": 204800, - "default_max_tokens": 131072, - "can_reason": true, - "reasoning_levels": [ - "low", - "medium", - "high" - ], - "default_reasoning_effort": "medium", - "supports_attachments": false - }, { "id": "minimax-m2.5", "name": "MiniMax M2.5", diff --git a/internal/providers/configs/openrouter.json b/internal/providers/configs/openrouter.json index b71f839c65abefe4f411d3f24676cdf1b7b2bf67..daee2ad4a43521f00c8cf42accfa58934e02448c 100644 --- a/internal/providers/configs/openrouter.json +++ b/internal/providers/configs/openrouter.json @@ -129,7 +129,7 @@ "cost_per_1m_in_cached": 3.75, "cost_per_1m_out_cached": 0.3, "context_window": 200000, - "default_max_tokens": 32000, + "default_max_tokens": 64000, "can_reason": true, "reasoning_levels": [ "low", @@ -586,10 +586,10 @@ { "id": "deepseek/deepseek-v3.2", "name": "DeepSeek: DeepSeek V3.2", - "cost_per_1m_in": 0.26, - "cost_per_1m_out": 0.38, + "cost_per_1m_in": 0.5, + "cost_per_1m_out": 1.5, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.13, + "cost_per_1m_out_cached": 0.25, "context_window": 163840, "default_max_tokens": 81920, "can_reason": true, @@ -607,9 +607,9 @@ "cost_per_1m_in": 0.27, "cost_per_1m_out": 0.41, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.27, + "cost_per_1m_out_cached": 0, "context_window": 163840, - "default_max_tokens": 81920, + "default_max_tokens": 32768, "can_reason": true, "reasoning_levels": [ "low", @@ -819,7 +819,7 @@ "cost_per_1m_in_cached": 0.08333, "cost_per_1m_out_cached": 0.05, "context_window": 1048576, - "default_max_tokens": 32768, + "default_max_tokens": 32767, "can_reason": true, "reasoning_levels": [ "low", @@ -886,10 +886,10 @@ { "id": "google/gemma-4-26b-a4b-it", "name": "Google: Gemma 4 26B A4B ", - "cost_per_1m_in": 0.1, + "cost_per_1m_in": 0.09, "cost_per_1m_out": 0.4, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.05, + "cost_per_1m_out_cached": 0.06, "context_window": 262144, "default_max_tokens": 131072, "can_reason": true, @@ -980,8 +980,8 @@ "cost_per_1m_out": 1.2, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.06, - "context_window": 262144, - "default_max_tokens": 72000, + "context_window": 256000, + "default_max_tokens": 40000, "can_reason": false, "supports_attachments": false }, @@ -993,7 +993,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 13107, + "default_max_tokens": 8192, "can_reason": false, "supports_attachments": false }, @@ -1033,31 +1033,13 @@ "can_reason": false, "supports_attachments": true }, - { - "id": "minimax/minimax-m1", - "name": "MiniMax: MiniMax M1", - "cost_per_1m_in": 0.44, - "cost_per_1m_out": 1.76, - "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 1000000, - "default_max_tokens": 20000, - "can_reason": true, - "reasoning_levels": [ - "low", - "medium", - "high" - ], - "default_reasoning_effort": "medium", - "supports_attachments": false - }, { "id": "minimax/minimax-m2", "name": "MiniMax: MiniMax M2", - "cost_per_1m_in": 0.3, - "cost_per_1m_out": 1.2, + "cost_per_1m_in": 0.255, + "cost_per_1m_out": 1, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.03, "context_window": 196608, "default_max_tokens": 98304, "can_reason": true, @@ -1090,12 +1072,12 @@ { "id": "minimax/minimax-m2.5", "name": "MiniMax: MiniMax M2.5", - "cost_per_1m_in": 0.3, - "cost_per_1m_out": 1.2, + "cost_per_1m_in": 0.6, + "cost_per_1m_out": 2.4, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.03, + "cost_per_1m_out_cached": 0.06, "context_window": 204800, - "default_max_tokens": 65550, + "default_max_tokens": 65536, "can_reason": true, "reasoning_levels": [ "low", @@ -1240,24 +1222,24 @@ { "id": "mistralai/ministral-3b-2512", "name": "Mistral: Ministral 3 3B 2512", - "cost_per_1m_in": 0.35, - "cost_per_1m_out": 0.35, + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.1, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.01, "context_window": 131072, - "default_max_tokens": 65536, + "default_max_tokens": 13107, "can_reason": false, "supports_attachments": true }, { "id": "mistralai/ministral-8b-2512", "name": "Mistral: Ministral 3 8B 2512", - "cost_per_1m_in": 0.3, - "cost_per_1m_out": 0.3, + "cost_per_1m_in": 0.15, + "cost_per_1m_out": 0.15, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.015, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 26214, "can_reason": false, "supports_attachments": true }, @@ -1426,12 +1408,12 @@ { "id": "moonshotai/kimi-k2-0905", "name": "MoonshotAI: Kimi K2 0905", - "cost_per_1m_in": 1, - "cost_per_1m_out": 3, + "cost_per_1m_in": 0.6, + "cost_per_1m_out": 2.5, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.5, + "cost_per_1m_out_cached": 0.6, "context_window": 262144, - "default_max_tokens": 8192, + "default_max_tokens": 4096, "can_reason": false, "supports_attachments": false }, @@ -1456,12 +1438,12 @@ { "id": "moonshotai/kimi-k2.5", "name": "MoonshotAI: Kimi K2.5", - "cost_per_1m_in": 0.45, - "cost_per_1m_out": 2.25, + "cost_per_1m_in": 0.445, + "cost_per_1m_out": 2, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.07, + "cost_per_1m_out_cached": 0.225, "context_window": 262144, - "default_max_tokens": 26214, + "default_max_tokens": 131072, "can_reason": true, "reasoning_levels": [ "low", @@ -2221,7 +2203,7 @@ "cost_per_1m_in": 0.15, "cost_per_1m_out": 0.6, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.014, "context_window": 131072, "default_max_tokens": 13107, "can_reason": true, @@ -2254,10 +2236,10 @@ { "id": "openai/gpt-oss-20b", "name": "OpenAI: gpt-oss-20b", - "cost_per_1m_in": 0.07, - "cost_per_1m_out": 0.3, + "cost_per_1m_in": 0.03, + "cost_per_1m_out": 0.14, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.035, + "cost_per_1m_out_cached": 0, "context_window": 131072, "default_max_tokens": 13107, "can_reason": true, @@ -2605,6 +2587,24 @@ "can_reason": false, "supports_attachments": false }, + { + "id": "qwen/qwen3-14b", + "name": "Qwen: Qwen3 14B", + "cost_per_1m_in": 0.2275, + "cost_per_1m_out": 0.91, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 131072, + "default_max_tokens": 4096, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false + }, { "id": "qwen/qwen3-235b-a22b", "name": "Qwen: Qwen3 235B A22B", @@ -2626,12 +2626,12 @@ { "id": "qwen/qwen3-235b-a22b-2507", "name": "Qwen: Qwen3 235B A22B Instruct 2507", - "cost_per_1m_in": 0.1, - "cost_per_1m_out": 0.1, + "cost_per_1m_in": 0.22, + "cost_per_1m_out": 0.88, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.1, + "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 8192, "can_reason": false, "supports_attachments": false }, @@ -2656,12 +2656,12 @@ { "id": "qwen/qwen3-30b-a3b", "name": "Qwen: Qwen3 30B A3B", - "cost_per_1m_in": 0.13, - "cost_per_1m_out": 0.52, + "cost_per_1m_in": 0.08, + "cost_per_1m_out": 0.28, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 131072, - "default_max_tokens": 4096, + "context_window": 40960, + "default_max_tokens": 20480, "can_reason": true, "reasoning_levels": [ "low", @@ -2677,9 +2677,9 @@ "cost_per_1m_in": 0.1, "cost_per_1m_out": 0.3, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.1, "context_window": 262144, - "default_max_tokens": 26214, + "default_max_tokens": 131072, "can_reason": false, "supports_attachments": false }, @@ -2788,12 +2788,12 @@ { "id": "qwen/qwen3-coder-next", "name": "Qwen: Qwen3 Coder Next", - "cost_per_1m_in": 0.15, - "cost_per_1m_out": 0.8, + "cost_per_1m_in": 0.5, + "cost_per_1m_out": 1.2, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.12, + "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 26214, "can_reason": false, "supports_attachments": false }, @@ -2842,12 +2842,12 @@ { "id": "qwen/qwen3-next-80b-a3b-instruct", "name": "Qwen: Qwen3 Next 80B A3B Instruct", - "cost_per_1m_in": 0.1, + "cost_per_1m_in": 0.09, "cost_per_1m_out": 1.1, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.07, + "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 26214, "can_reason": false, "supports_attachments": false }, @@ -2914,12 +2914,12 @@ { "id": "qwen/qwen3-vl-30b-a3b-instruct", "name": "Qwen: Qwen3 VL 30B A3B Instruct", - "cost_per_1m_in": 0.29, - "cost_per_1m_out": 1, + "cost_per_1m_in": 0.13, + "cost_per_1m_out": 0.52, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 262144, - "default_max_tokens": 131072, + "context_window": 131072, + "default_max_tokens": 16384, "can_reason": false, "supports_attachments": true }, @@ -3022,10 +3022,10 @@ { "id": "qwen/qwen3.5-122b-a10b", "name": "Qwen: Qwen3.5-122B-A10B", - "cost_per_1m_in": 0.26, - "cost_per_1m_out": 2.08, + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 2.4, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.3, "context_window": 262144, "default_max_tokens": 32768, "can_reason": true, @@ -3040,10 +3040,10 @@ { "id": "qwen/qwen3.5-27b", "name": "Qwen: Qwen3.5-27B", - "cost_per_1m_in": 0.27, - "cost_per_1m_out": 2.16, + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 2.4, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.27, + "cost_per_1m_out_cached": 0, "context_window": 262144, "default_max_tokens": 32768, "can_reason": true, @@ -3058,10 +3058,10 @@ { "id": "qwen/qwen3.5-35b-a3b", "name": "Qwen: Qwen3.5-35B-A3B", - "cost_per_1m_in": 0.23, - "cost_per_1m_out": 1.8, + "cost_per_1m_in": 0.2, + "cost_per_1m_out": 1, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.1, "context_window": 262144, "default_max_tokens": 131072, "can_reason": true, @@ -3080,8 +3080,8 @@ "cost_per_1m_out": 0.15, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 262144, - "default_max_tokens": 26214, + "context_window": 256000, + "default_max_tokens": 16384, "can_reason": true, "reasoning_levels": [ "low", @@ -3145,9 +3145,9 @@ "cost_per_1m_in": 0.1, "cost_per_1m_out": 0.3, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 262144, - "default_max_tokens": 32768, + "cost_per_1m_out_cached": 0.02, + "context_window": 256000, + "default_max_tokens": 128000, "can_reason": true, "reasoning_levels": [ "low", @@ -3309,7 +3309,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.11, "context_window": 131072, - "default_max_tokens": 49152, + "default_max_tokens": 48000, "can_reason": true, "reasoning_levels": [ "low", @@ -3412,12 +3412,12 @@ { "id": "z-ai/glm-4.7", "name": "Z.ai: GLM 4.7", - "cost_per_1m_in": 0.45, - "cost_per_1m_out": 2.2, + "cost_per_1m_in": 0.54, + "cost_per_1m_out": 1.98, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.11, + "cost_per_1m_out_cached": 0.099, "context_window": 204800, - "default_max_tokens": 102400, + "default_max_tokens": 65536, "can_reason": true, "reasoning_levels": [ "low", @@ -3430,12 +3430,12 @@ { "id": "z-ai/glm-4.7-flash", "name": "Z.ai: GLM 4.7 Flash", - "cost_per_1m_in": 0.1, - "cost_per_1m_out": 0.43, + "cost_per_1m_in": 0.06, + "cost_per_1m_out": 0.4, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.01, "context_window": 202752, - "default_max_tokens": 101376, + "default_max_tokens": 20275, "can_reason": true, "reasoning_levels": [ "low", diff --git a/internal/providers/configs/venice.json b/internal/providers/configs/venice.json index acf9c73842dca3badb437901e0fad7f5894627c5..c74301aad382f4db23648d71cf5cd8130313d0dd 100644 --- a/internal/providers/configs/venice.json +++ b/internal/providers/configs/venice.json @@ -517,6 +517,24 @@ "default_reasoning_effort": "medium", "supports_attachments": true }, + { + "id": "kimi-k2-6", + "name": "Kimi K2.6", + "cost_per_1m_in": 0.56, + "cost_per_1m_out": 3.5, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 256000, + "default_max_tokens": 65536, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": true + }, { "id": "llama-3.2-3b", "name": "Llama 3.2 3B", diff --git a/internal/providers/configs/vercel.json b/internal/providers/configs/vercel.json index d29b59b39b8745e194394a8cfc4dcf6d6917e2b1..711efd8dbf21ad2adec564ae97ac3e2246250f1f 100644 --- a/internal/providers/configs/vercel.json +++ b/internal/providers/configs/vercel.json @@ -2155,6 +2155,24 @@ "default_reasoning_effort": "medium", "supports_attachments": true }, + { + "id": "alibaba/qwen-3.6-max-preview", + "name": "Qwen 3.6 Max Preview", + "cost_per_1m_in": 1.3, + "cost_per_1m_out": 7.8, + "cost_per_1m_in_cached": 0.26, + "cost_per_1m_out_cached": 1.625, + "context_window": 240000, + "default_max_tokens": 8000, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": true + }, { "id": "alibaba/qwen3.6-plus", "name": "Qwen 3.6 Plus",