diff --git a/internal/providers/configs/aihubmix.json b/internal/providers/configs/aihubmix.json index 0e124f99c6ca428a95f1e8d887cc5a920f9dc433..0787344ef4464ae669034371ce97e552d403acf2 100644 --- a/internal/providers/configs/aihubmix.json +++ b/internal/providers/configs/aihubmix.json @@ -2587,6 +2587,25 @@ "supports_attachments": false, "options": {} }, + { + "id": "minimax-m2.7", + "name": "MiniMax M2.7", + "cost_per_1m_in": 0.2958, + "cost_per_1m_out": 1.1832, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0.05916, + "context_window": 200000, + "default_max_tokens": 20000, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false, + "options": {} + }, { "id": "mistral-large-3", "name": "Mistral Large 3", diff --git a/internal/providers/configs/huggingface.json b/internal/providers/configs/huggingface.json index 9d4189a85dc7fb887bc7e572944703b65805a999..1caa40210683531cfe294825642f05ec6144c698 100644 --- a/internal/providers/configs/huggingface.json +++ b/internal/providers/configs/huggingface.json @@ -111,19 +111,6 @@ "supports_attachments": false, "options": {} }, - { - "id": "deepseek-ai/DeepSeek-V3.2:fireworks-ai", - "name": "deepseek-ai/DeepSeek-V3.2 (fireworks-ai)", - "cost_per_1m_in": 0, - "cost_per_1m_out": 0, - "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 163840, - "default_max_tokens": 8192, - "can_reason": false, - "supports_attachments": false, - "options": {} - }, { "id": "meta-llama/Llama-3.3-70B-Instruct:groq", "name": "meta-llama/Llama-3.3-70B-Instruct (groq)", diff --git a/internal/providers/configs/openrouter.json b/internal/providers/configs/openrouter.json index 8a5b27bef09326ea00b72befcddbd9227d501972..37044436ba1972b055a4102bff33a7ca60a830dd 100644 --- a/internal/providers/configs/openrouter.json +++ b/internal/providers/configs/openrouter.json @@ -566,12 +566,12 @@ { "id": "deepseek/deepseek-chat-v3.1", "name": "DeepSeek: DeepSeek V3.1", - "cost_per_1m_in": 0.2, - "cost_per_1m_out": 0.8, + "cost_per_1m_in": 0.21, + "cost_per_1m_out": 0.79, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.1, + "cost_per_1m_out_cached": 0.13, "context_window": 163840, - "default_max_tokens": 32768, + "default_max_tokens": 16384, "can_reason": true, "reasoning_levels": [ "low", @@ -604,12 +604,12 @@ { "id": "deepseek/deepseek-v3.2", "name": "DeepSeek: DeepSeek V3.2", - "cost_per_1m_in": 0.26, - "cost_per_1m_out": 0.38, + "cost_per_1m_in": 0.56, + "cost_per_1m_out": 1.68, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.13, + "cost_per_1m_out_cached": 0, "context_window": 163840, - "default_max_tokens": 81920, + "default_max_tokens": 32768, "can_reason": true, "reasoning_levels": [ "low", @@ -626,9 +626,9 @@ "cost_per_1m_in": 0.27, "cost_per_1m_out": 0.41, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.27, + "cost_per_1m_out_cached": 0, "context_window": 163840, - "default_max_tokens": 81920, + "default_max_tokens": 32768, "can_reason": true, "reasoning_levels": [ "low", @@ -661,12 +661,12 @@ { "id": "deepseek/deepseek-r1-0528", "name": "DeepSeek: R1 0528", - "cost_per_1m_in": 0.45, - "cost_per_1m_out": 2.15, + "cost_per_1m_in": 0.7, + "cost_per_1m_out": 2.5, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.225, + "cost_per_1m_out_cached": 0.35, "context_window": 163840, - "default_max_tokens": 32768, + "default_max_tokens": 16384, "can_reason": true, "reasoning_levels": [ "low", @@ -838,7 +838,7 @@ "cost_per_1m_in_cached": 0.08333, "cost_per_1m_out_cached": 0.05, "context_window": 1048576, - "default_max_tokens": 32767, + "default_max_tokens": 32768, "can_reason": true, "reasoning_levels": [ "low", @@ -1004,7 +1004,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 13107, + "default_max_tokens": 8192, "can_reason": false, "supports_attachments": false, "options": {} @@ -1012,12 +1012,12 @@ { "id": "meta-llama/llama-3.3-70b-instruct", "name": "Meta: Llama 3.3 70B Instruct", - "cost_per_1m_in": 0.13, - "cost_per_1m_out": 0.4, + "cost_per_1m_in": 0.25, + "cost_per_1m_out": 0.75, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 64000, + "default_max_tokens": 13107, "can_reason": false, "supports_attachments": false, "options": {} @@ -1140,8 +1140,8 @@ { "id": "minimax/minimax-m2.7", "name": "MiniMax: MiniMax M2.7", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 2.4, + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 1.2, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.06, "context_window": 204800, @@ -1279,9 +1279,9 @@ "cost_per_1m_in": 0.15, "cost_per_1m_out": 0.15, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.015, + "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 26214, + "default_max_tokens": 131072, "can_reason": false, "supports_attachments": true, "options": {} @@ -1477,12 +1477,12 @@ { "id": "moonshotai/kimi-k2", "name": "MoonshotAI: Kimi K2 0711", - "cost_per_1m_in": 0.57, - "cost_per_1m_out": 2.3, + "cost_per_1m_in": 0.6, + "cost_per_1m_out": 2.5, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.15, "context_window": 131072, - "default_max_tokens": 65536, + "default_max_tokens": 13107, "can_reason": false, "supports_attachments": false, "options": {} @@ -1490,12 +1490,12 @@ { "id": "moonshotai/kimi-k2-0905", "name": "MoonshotAI: Kimi K2 0905", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 2.5, + "cost_per_1m_in": 1, + "cost_per_1m_out": 3, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.15, + "cost_per_1m_out_cached": 0.5, "context_window": 262144, - "default_max_tokens": 26214, + "default_max_tokens": 8192, "can_reason": false, "supports_attachments": false, "options": {} @@ -1506,7 +1506,7 @@ "cost_per_1m_in": 0.6, "cost_per_1m_out": 2.5, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.15, + "cost_per_1m_out_cached": 0, "context_window": 262144, "default_max_tokens": 131072, "can_reason": true, @@ -1776,7 +1776,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.1, "context_window": 1047576, - "default_max_tokens": 16384, + "default_max_tokens": 104757, "can_reason": false, "supports_attachments": true, "options": {} @@ -2302,11 +2302,11 @@ "id": "openai/gpt-oss-120b", "name": "OpenAI: gpt-oss-120b", "cost_per_1m_in": 0.05, - "cost_per_1m_out": 0.45, + "cost_per_1m_out": 0.25, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.025, + "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 32768, + "default_max_tokens": 16384, "can_reason": true, "reasoning_levels": [ "low", @@ -2712,25 +2712,6 @@ "supports_attachments": false, "options": {} }, - { - "id": "qwen/qwen3-14b", - "name": "Qwen: Qwen3 14B", - "cost_per_1m_in": 0.2275, - "cost_per_1m_out": 0.91, - "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 131072, - "default_max_tokens": 4096, - "can_reason": true, - "reasoning_levels": [ - "low", - "medium", - "high" - ], - "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} - }, { "id": "qwen/qwen3-235b-a22b", "name": "Qwen: Qwen3 235B A22B", @@ -2753,12 +2734,12 @@ { "id": "qwen/qwen3-235b-a22b-2507", "name": "Qwen: Qwen3 235B A22B Instruct 2507", - "cost_per_1m_in": 0.22, - "cost_per_1m_out": 0.88, + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.1, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 8192, + "default_max_tokens": 131072, "can_reason": false, "supports_attachments": false, "options": {} @@ -2766,12 +2747,12 @@ { "id": "qwen/qwen3-235b-a22b-thinking-2507", "name": "Qwen: Qwen3 235B A22B Thinking 2507", - "cost_per_1m_in": 0.3, - "cost_per_1m_out": 3, + "cost_per_1m_in": 0.1495, + "cost_per_1m_out": 1.495, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 16384, + "default_max_tokens": 13107, "can_reason": true, "reasoning_levels": [ "low", @@ -2785,12 +2766,12 @@ { "id": "qwen/qwen3-30b-a3b", "name": "Qwen: Qwen3 30B A3B", - "cost_per_1m_in": 0.13, - "cost_per_1m_out": 0.52, + "cost_per_1m_in": 0.08, + "cost_per_1m_out": 0.28, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 131072, - "default_max_tokens": 4096, + "context_window": 40960, + "default_max_tokens": 20480, "can_reason": true, "reasoning_levels": [ "low", @@ -2894,11 +2875,11 @@ "id": "qwen/qwen3-coder-30b-a3b-instruct", "name": "Qwen: Qwen3 Coder 30B A3B Instruct", "cost_per_1m_in": 0.07, - "cost_per_1m_out": 0.28, + "cost_per_1m_out": 0.27, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 262144, - "default_max_tokens": 131072, + "context_window": 160000, + "default_max_tokens": 16384, "can_reason": false, "supports_attachments": false, "options": {} @@ -2907,11 +2888,11 @@ "id": "qwen/qwen3-coder", "name": "Qwen: Qwen3 Coder 480B A35B", "cost_per_1m_in": 0.22, - "cost_per_1m_out": 1, + "cost_per_1m_out": 1.8, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.022, + "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 26214, + "default_max_tokens": 32768, "can_reason": false, "supports_attachments": false, "options": {} @@ -2945,12 +2926,12 @@ { "id": "qwen/qwen3-coder-next", "name": "Qwen: Qwen3 Coder Next", - "cost_per_1m_in": 0.18, - "cost_per_1m_out": 1.35, + "cost_per_1m_in": 0.12, + "cost_per_1m_out": 0.75, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.18, + "cost_per_1m_out_cached": 0.06, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 32768, "can_reason": false, "supports_attachments": false, "options": {} @@ -3003,8 +2984,8 @@ { "id": "qwen/qwen3-next-80b-a3b-instruct", "name": "Qwen: Qwen3 Next 80B A3B Instruct", - "cost_per_1m_in": 0.1, - "cost_per_1m_out": 1.1, + "cost_per_1m_in": 0.15, + "cost_per_1m_out": 1.2, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, @@ -3093,12 +3074,12 @@ { "id": "qwen/qwen3-vl-30b-a3b-thinking", "name": "Qwen: Qwen3 VL 30B A3B Thinking", - "cost_per_1m_in": 0.29, + "cost_per_1m_in": 0.2, "cost_per_1m_out": 1, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 262144, - "default_max_tokens": 131072, + "context_window": 131072, + "default_max_tokens": 16384, "can_reason": true, "reasoning_levels": [ "low", @@ -3214,10 +3195,10 @@ { "id": "qwen/qwen3.5-27b", "name": "Qwen: Qwen3.5-27B", - "cost_per_1m_in": 0.27, - "cost_per_1m_out": 2.16, + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 2.4, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.27, + "cost_per_1m_out_cached": 0, "context_window": 262144, "default_max_tokens": 32768, "can_reason": true, @@ -3233,12 +3214,12 @@ { "id": "qwen/qwen3.5-35b-a3b", "name": "Qwen: Qwen3.5-35B-A3B", - "cost_per_1m_in": 0.25, - "cost_per_1m_out": 1, + "cost_per_1m_in": 0.225, + "cost_per_1m_out": 1.8, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.225, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 32768, "can_reason": true, "reasoning_levels": [ "low", @@ -3252,12 +3233,12 @@ { "id": "qwen/qwen3.5-9b", "name": "Qwen: Qwen3.5-9B", - "cost_per_1m_in": 0.1, + "cost_per_1m_in": 0.05, "cost_per_1m_out": 0.15, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 262144, - "default_max_tokens": 26214, + "context_window": 256000, + "default_max_tokens": 25600, "can_reason": true, "reasoning_levels": [ "low", @@ -3513,12 +3494,12 @@ { "id": "z-ai/glm-4.5-air", "name": "Z.ai: GLM 4.5 Air", - "cost_per_1m_in": 0.14, - "cost_per_1m_out": 0.86, + "cost_per_1m_in": 0.13, + "cost_per_1m_out": 0.85, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.025, "context_window": 131072, - "default_max_tokens": 65536, + "default_max_tokens": 49152, "can_reason": true, "reasoning_levels": [ "low", @@ -3608,12 +3589,12 @@ { "id": "z-ai/glm-4.7", "name": "Z.ai: GLM 4.7", - "cost_per_1m_in": 0.54, - "cost_per_1m_out": 1.98, + "cost_per_1m_in": 0.45, + "cost_per_1m_out": 2.2, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.099, + "cost_per_1m_out_cached": 0.11, "context_window": 204800, - "default_max_tokens": 65536, + "default_max_tokens": 102400, "can_reason": true, "reasoning_levels": [ "low",