diff --git a/internal/providers/configs/aihubmix.json b/internal/providers/configs/aihubmix.json index d5988106597ae227c2b16928528412fdabe5e531..d485b262ac6a97293ec8b5fa60764774ab8641be 100644 --- a/internal/providers/configs/aihubmix.json +++ b/internal/providers/configs/aihubmix.json @@ -658,6 +658,63 @@ "supports_attachments": false, "options": {} }, + { + "id": "coding-minimax-m2.7", + "name": "Coding MiniMax M2.7", + "cost_per_1m_in": 0.2, + "cost_per_1m_out": 0.2, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 204800, + "default_max_tokens": 13100, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false, + "options": {} + }, + { + "id": "coding-minimax-m2.7-free", + "name": "Coding MiniMax M2.7 (free)", + "cost_per_1m_in": 0, + "cost_per_1m_out": 0, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 204800, + "default_max_tokens": 13100, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false, + "options": {} + }, + { + "id": "coding-minimax-m2.7-highspeed", + "name": "Coding MiniMax M2.7 Highspeed", + "cost_per_1m_in": 0.2, + "cost_per_1m_out": 0.2, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 204800, + "default_max_tokens": 13100, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false, + "options": {} + }, { "id": "deepseek-math-v2", "name": "DeepSeek Math V2", @@ -2014,8 +2071,8 @@ "cost_per_1m_out": 4.5, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.075, - "context_window": 1047576, - "default_max_tokens": 32768, + "context_window": 400000, + "default_max_tokens": 128000, "can_reason": false, "supports_attachments": true, "options": {} @@ -2466,6 +2523,25 @@ "supports_attachments": false, "options": {} }, + { + "id": "minimax-m2.5-free", + "name": "MiniMax M2.5 (free)", + "cost_per_1m_in": 0, + "cost_per_1m_out": 0, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 204800, + "default_max_tokens": 20480, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false, + "options": {} + }, { "id": "minimax-m2.5-highspeed", "name": "MiniMax M2.5 Highspeed", diff --git a/internal/providers/configs/huggingface.json b/internal/providers/configs/huggingface.json index 8cb6ca8360a6538f095648587892e1c1965ec738..38daf7e3ef1faf6c3999b89dd2c7de35891a68e1 100644 --- a/internal/providers/configs/huggingface.json +++ b/internal/providers/configs/huggingface.json @@ -20,19 +20,6 @@ "supports_attachments": false, "options": {} }, - { - "id": "MiniMaxAI/MiniMax-M2.5:fireworks-ai", - "name": "MiniMaxAI/MiniMax-M2.5 (fireworks-ai)", - "cost_per_1m_in": 0, - "cost_per_1m_out": 0, - "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 204800, - "default_max_tokens": 8192, - "can_reason": false, - "supports_attachments": false, - "options": {} - }, { "id": "Qwen/Qwen3-32B:groq", "name": "Qwen/Qwen3-32B (groq)", @@ -111,19 +98,6 @@ "supports_attachments": false, "options": {} }, - { - "id": "deepseek-ai/DeepSeek-V3.2:fireworks-ai", - "name": "deepseek-ai/DeepSeek-V3.2 (fireworks-ai)", - "cost_per_1m_in": 0, - "cost_per_1m_out": 0, - "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 163840, - "default_max_tokens": 8192, - "can_reason": false, - "supports_attachments": false, - "options": {} - }, { "id": "meta-llama/Llama-3.3-70B-Instruct:groq", "name": "meta-llama/Llama-3.3-70B-Instruct (groq)", diff --git a/internal/providers/configs/ionet.json b/internal/providers/configs/ionet.json index 8b5b5b314c97b48b8beb8cc0ae0c8ded802867ba..4016b3760d7e6b6990d130a785b95725e117e43c 100644 --- a/internal/providers/configs/ionet.json +++ b/internal/providers/configs/ionet.json @@ -84,6 +84,19 @@ "supports_attachments": false, "options": {} }, + { + "id": "moonshotai/Kimi-K2.5", + "name": "MoonshotAI: Kimi K2.5", + "cost_per_1m_in": 0.55, + "cost_per_1m_out": 3, + "cost_per_1m_in_cached": 0.275, + "cost_per_1m_out_cached": 1.1, + "context_window": 262144, + "default_max_tokens": 26214, + "can_reason": false, + "supports_attachments": false, + "options": {} + }, { "id": "Qwen/Qwen3-Next-80B-A3B-Instruct", "name": "Qwen: Qwen3 Next 80B A3B Instruct", diff --git a/internal/providers/configs/openrouter.json b/internal/providers/configs/openrouter.json index 4cd866df68afebb726457cb069a03a2ed0157066..2352f2835d5b7383f5b55480db94e0e402f00e0f 100644 --- a/internal/providers/configs/openrouter.json +++ b/internal/providers/configs/openrouter.json @@ -547,31 +547,25 @@ { "id": "deepseek/deepseek-chat-v3-0324", "name": "DeepSeek: DeepSeek V3 0324", - "cost_per_1m_in": 0.77, - "cost_per_1m_out": 0.77, + "cost_per_1m_in": 0.27, + "cost_per_1m_out": 1.12, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.135, "context_window": 163840, - "default_max_tokens": 65536, - "can_reason": true, - "reasoning_levels": [ - "low", - "medium", - "high" - ], - "default_reasoning_effort": "medium", + "default_max_tokens": 81920, + "can_reason": false, "supports_attachments": false, "options": {} }, { "id": "deepseek/deepseek-chat-v3.1", "name": "DeepSeek: DeepSeek V3.1", - "cost_per_1m_in": 0.2, - "cost_per_1m_out": 0.8, + "cost_per_1m_in": 0.21, + "cost_per_1m_out": 0.79, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.1, + "cost_per_1m_out_cached": 0.13, "context_window": 163840, - "default_max_tokens": 32768, + "default_max_tokens": 16384, "can_reason": true, "reasoning_levels": [ "low", @@ -609,7 +603,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.13, "context_window": 163840, - "default_max_tokens": 81920, + "default_max_tokens": 16384, "can_reason": true, "reasoning_levels": [ "low", @@ -626,9 +620,9 @@ "cost_per_1m_in": 0.27, "cost_per_1m_out": 0.41, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.27, "context_window": 163840, - "default_max_tokens": 32768, + "default_max_tokens": 81920, "can_reason": true, "reasoning_levels": [ "low", @@ -925,70 +919,6 @@ "supports_attachments": true, "options": {} }, - { - "id": "google/gemma-3-27b-it", - "name": "Google: Gemma 3 27B", - "cost_per_1m_in": 0.03, - "cost_per_1m_out": 0.11, - "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.015, - "context_window": 128000, - "default_max_tokens": 32768, - "can_reason": false, - "supports_attachments": true, - "options": {} - }, - { - "id": "google/gemma-3-27b-it:free", - "name": "Google: Gemma 3 27B (free)", - "cost_per_1m_in": 0, - "cost_per_1m_out": 0, - "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 128000, - "default_max_tokens": 64000, - "can_reason": false, - "supports_attachments": true, - "options": {} - }, - { - "id": "openrouter/healer-alpha", - "name": "Healer Alpha", - "cost_per_1m_in": 0, - "cost_per_1m_out": 0, - "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 262144, - "default_max_tokens": 16000, - "can_reason": true, - "reasoning_levels": [ - "low", - "medium", - "high" - ], - "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} - }, - { - "id": "openrouter/hunter-alpha", - "name": "Hunter Alpha", - "cost_per_1m_in": 0, - "cost_per_1m_out": 0, - "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 1048576, - "default_max_tokens": 16000, - "can_reason": true, - "reasoning_levels": [ - "low", - "medium", - "high" - ], - "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} - }, { "id": "inception/mercury", "name": "Inception: Mercury", @@ -1076,8 +1006,8 @@ { "id": "meta-llama/llama-3.3-70b-instruct", "name": "Meta: Llama 3.3 70B Instruct", - "cost_per_1m_in": 0.13, - "cost_per_1m_out": 0.4, + "cost_per_1m_in": 0.25, + "cost_per_1m_out": 0.75, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, @@ -1093,8 +1023,8 @@ "cost_per_1m_out": 0, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 128000, - "default_max_tokens": 64000, + "context_window": 65536, + "default_max_tokens": 6553, "can_reason": false, "supports_attachments": false, "options": {} @@ -1166,12 +1096,12 @@ { "id": "minimax/minimax-m2.5", "name": "MiniMax: MiniMax M2.5", - "cost_per_1m_in": 0.3, - "cost_per_1m_out": 1.2, + "cost_per_1m_in": 0.6, + "cost_per_1m_out": 2.4, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.03, + "cost_per_1m_out_cached": 0.06, "context_window": 204800, - "default_max_tokens": 65550, + "default_max_tokens": 65536, "can_reason": true, "reasoning_levels": [ "low", @@ -1201,6 +1131,25 @@ "supports_attachments": false, "options": {} }, + { + "id": "minimax/minimax-m2.7", + "name": "MiniMax: MiniMax M2.7", + "cost_per_1m_in": 0.6, + "cost_per_1m_out": 2.4, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0.06, + "context_window": 204800, + "default_max_tokens": 65536, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false, + "options": {} + }, { "id": "mistralai/mistral-large", "name": "Mistral Large", @@ -1311,9 +1260,9 @@ "cost_per_1m_in": 0.1, "cost_per_1m_out": 0.1, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.01, "context_window": 131072, - "default_max_tokens": 65536, + "default_max_tokens": 13107, "can_reason": false, "supports_attachments": true, "options": {} @@ -1324,9 +1273,9 @@ "cost_per_1m_in": 0.15, "cost_per_1m_out": 0.15, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.015, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 26214, "can_reason": false, "supports_attachments": true, "options": {} @@ -1370,6 +1319,19 @@ "supports_attachments": true, "options": {} }, + { + "id": "mistralai/mistral-nemo", + "name": "Mistral: Mistral Nemo", + "cost_per_1m_in": 0.15, + "cost_per_1m_out": 0.15, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0.015, + "context_window": 131072, + "default_max_tokens": 13107, + "can_reason": false, + "supports_attachments": false, + "options": {} + }, { "id": "mistralai/mistral-small-24b-instruct-2501", "name": "Mistral: Mistral Small 3", @@ -1399,12 +1361,12 @@ { "id": "mistralai/mistral-small-3.2-24b-instruct", "name": "Mistral: Mistral Small 3.2 24B", - "cost_per_1m_in": 0.06, - "cost_per_1m_out": 0.18, + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.3, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.03, + "cost_per_1m_out_cached": 0.01, "context_window": 131072, - "default_max_tokens": 65536, + "default_max_tokens": 13107, "can_reason": false, "supports_attachments": true, "options": {} @@ -1522,12 +1484,12 @@ { "id": "moonshotai/kimi-k2-0905", "name": "MoonshotAI: Kimi K2 0905", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 2.5, + "cost_per_1m_in": 1, + "cost_per_1m_out": 3, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.6, + "cost_per_1m_out_cached": 0.5, "context_window": 262144, - "default_max_tokens": 16384, + "default_max_tokens": 8192, "can_reason": false, "supports_attachments": false, "options": {} @@ -1554,12 +1516,12 @@ { "id": "moonshotai/kimi-k2.5", "name": "MoonshotAI: Kimi K2.5", - "cost_per_1m_in": 0.45, - "cost_per_1m_out": 2.2, + "cost_per_1m_in": 0.6, + "cost_per_1m_out": 2.8, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.225, + "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 32767, + "default_max_tokens": 131072, "can_reason": true, "reasoning_levels": [ "low", @@ -1776,7 +1738,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.5, "context_window": 1047576, - "default_max_tokens": 16384, + "default_max_tokens": 104757, "can_reason": false, "supports_attachments": true, "options": {} @@ -1813,7 +1775,7 @@ "cost_per_1m_in": 2.5, "cost_per_1m_out": 10, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 1.25, "context_window": 128000, "default_max_tokens": 8192, "can_reason": false, @@ -2314,12 +2276,12 @@ { "id": "openai/gpt-oss-120b", "name": "OpenAI: gpt-oss-120b", - "cost_per_1m_in": 0.1, - "cost_per_1m_out": 0.49, + "cost_per_1m_in": 0.05, + "cost_per_1m_out": 0.45, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.025, "context_window": 131072, - "default_max_tokens": 13107, + "default_max_tokens": 32768, "can_reason": true, "reasoning_levels": [ "low", @@ -2747,8 +2709,8 @@ { "id": "qwen/qwen3-235b-a22b-2507", "name": "Qwen: Qwen3 235B A22B Instruct 2507", - "cost_per_1m_in": 0.2, - "cost_per_1m_out": 0.8, + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.1, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, @@ -2779,12 +2741,12 @@ { "id": "qwen/qwen3-30b-a3b", "name": "Qwen: Qwen3 30B A3B", - "cost_per_1m_in": 0.15, - "cost_per_1m_out": 0.6, + "cost_per_1m_in": 0.13, + "cost_per_1m_out": 0.52, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.15, + "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 65536, + "default_max_tokens": 4096, "can_reason": true, "reasoning_levels": [ "low", @@ -2808,6 +2770,25 @@ "supports_attachments": false, "options": {} }, + { + "id": "qwen/qwen3-30b-a3b-thinking-2507", + "name": "Qwen: Qwen3 30B A3B Thinking 2507", + "cost_per_1m_in": 0.08, + "cost_per_1m_out": 0.4, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0.08, + "context_window": 131072, + "default_max_tokens": 65536, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false, + "options": {} + }, { "id": "qwen/qwen3-32b", "name": "Qwen: Qwen3 32B", @@ -2868,12 +2849,12 @@ { "id": "qwen/qwen3-coder-30b-a3b-instruct", "name": "Qwen: Qwen3 Coder 30B A3B Instruct", - "cost_per_1m_in": 0.07, - "cost_per_1m_out": 0.27, + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.3, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 160000, - "default_max_tokens": 16384, + "context_window": 262144, + "default_max_tokens": 26214, "can_reason": false, "supports_attachments": false, "options": {} @@ -2920,12 +2901,12 @@ { "id": "qwen/qwen3-coder-next", "name": "Qwen: Qwen3 Coder Next", - "cost_per_1m_in": 0.18, - "cost_per_1m_out": 1.35, + "cost_per_1m_in": 0.12, + "cost_per_1m_out": 0.75, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.18, + "cost_per_1m_out_cached": 0.06, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 32768, "can_reason": false, "supports_attachments": false, "options": {} @@ -2978,12 +2959,12 @@ { "id": "qwen/qwen3-next-80b-a3b-instruct", "name": "Qwen: Qwen3 Next 80B A3B Instruct", - "cost_per_1m_in": 0.1, + "cost_per_1m_in": 0.09, "cost_per_1m_out": 1.1, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 26214, "can_reason": false, "supports_attachments": false, "options": {} @@ -3004,8 +2985,8 @@ { "id": "qwen/qwen3-next-80b-a3b-thinking", "name": "Qwen: Qwen3 Next 80B A3B Thinking", - "cost_per_1m_in": 0.15, - "cost_per_1m_out": 1.2, + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 0.3, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, @@ -3036,12 +3017,12 @@ { "id": "qwen/qwen3-vl-235b-a22b-thinking", "name": "Qwen: Qwen3 VL 235B A22B Thinking", - "cost_per_1m_in": 0.45, - "cost_per_1m_out": 3.5, + "cost_per_1m_in": 0.26, + "cost_per_1m_out": 2.6, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 262144, - "default_max_tokens": 131072, + "context_window": 131072, + "default_max_tokens": 16384, "can_reason": true, "reasoning_levels": [ "low", @@ -3068,12 +3049,12 @@ { "id": "qwen/qwen3-vl-30b-a3b-thinking", "name": "Qwen: Qwen3 VL 30B A3B Thinking", - "cost_per_1m_in": 0.29, + "cost_per_1m_in": 0.2, "cost_per_1m_out": 1, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 262144, - "default_max_tokens": 131072, + "context_window": 131072, + "default_max_tokens": 16384, "can_reason": true, "reasoning_levels": [ "low", @@ -3170,10 +3151,10 @@ { "id": "qwen/qwen3.5-122b-a10b", "name": "Qwen: Qwen3.5-122B-A10B", - "cost_per_1m_in": 0.3, - "cost_per_1m_out": 2.4, + "cost_per_1m_in": 0.4, + "cost_per_1m_out": 3.2, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.3, + "cost_per_1m_out_cached": 0, "context_window": 262144, "default_max_tokens": 32768, "can_reason": true, @@ -3189,8 +3170,8 @@ { "id": "qwen/qwen3.5-27b", "name": "Qwen: Qwen3.5-27B", - "cost_per_1m_in": 0.195, - "cost_per_1m_out": 1.56, + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 2.4, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, @@ -3208,12 +3189,12 @@ { "id": "qwen/qwen3.5-35b-a3b", "name": "Qwen: Qwen3.5-35B-A3B", - "cost_per_1m_in": 0.225, - "cost_per_1m_out": 1.8, + "cost_per_1m_in": 0.25, + "cost_per_1m_out": 1, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.225, + "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 32768, + "default_max_tokens": 131072, "can_reason": true, "reasoning_levels": [ "low", @@ -3399,10 +3380,10 @@ { "id": "xiaomi/mimo-v2-flash", "name": "Xiaomi: MiMo-V2-Flash", - "cost_per_1m_in": 0.1, - "cost_per_1m_out": 0.3, + "cost_per_1m_in": 0.09, + "cost_per_1m_out": 0.29, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.01, + "cost_per_1m_out_cached": 0.045, "context_window": 262144, "default_max_tokens": 32768, "can_reason": true, @@ -3415,6 +3396,44 @@ "supports_attachments": false, "options": {} }, + { + "id": "xiaomi/mimo-v2-omni", + "name": "Xiaomi: MiMo-V2-Omni", + "cost_per_1m_in": 0.4, + "cost_per_1m_out": 2, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0.08, + "context_window": 262144, + "default_max_tokens": 32768, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": true, + "options": {} + }, + { + "id": "xiaomi/mimo-v2-pro", + "name": "Xiaomi: MiMo-V2-Pro", + "cost_per_1m_in": 1, + "cost_per_1m_out": 3, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0.2, + "context_window": 1048576, + "default_max_tokens": 65536, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false, + "options": {} + }, { "id": "z-ai/glm-4-32b", "name": "Z.ai: GLM 4 32B ", @@ -3450,12 +3469,12 @@ { "id": "z-ai/glm-4.5-air", "name": "Z.ai: GLM 4.5 Air", - "cost_per_1m_in": 0.14, - "cost_per_1m_out": 0.86, + "cost_per_1m_in": 0.2, + "cost_per_1m_out": 1.2, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 65536, + "default_max_tokens": 13107, "can_reason": true, "reasoning_levels": [ "low", @@ -3545,12 +3564,12 @@ { "id": "z-ai/glm-4.7", "name": "Z.ai: GLM 4.7", - "cost_per_1m_in": 0.54, - "cost_per_1m_out": 1.98, + "cost_per_1m_in": 0.44, + "cost_per_1m_out": 2, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.099, - "context_window": 204800, - "default_max_tokens": 65536, + "cost_per_1m_out_cached": 0.22, + "context_window": 202752, + "default_max_tokens": 20275, "can_reason": true, "reasoning_levels": [ "low", diff --git a/internal/providers/configs/synthetic.json b/internal/providers/configs/synthetic.json index e2a209aee3389bc04b1a55f867b304be837d342b..0ef0c5759f86acdc35e0579e893951e555c13b71 100644 --- a/internal/providers/configs/synthetic.json +++ b/internal/providers/configs/synthetic.json @@ -80,10 +80,10 @@ { "id": "hf:zai-org/GLM-4.7-Flash", "name": "GLM 4.7 Flash", - "cost_per_1m_in": 0.06, - "cost_per_1m_out": 0.4, - "cost_per_1m_in_cached": 0.06, - "cost_per_1m_out_cached": 0.06, + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.5, + "cost_per_1m_in_cached": 0.1, + "cost_per_1m_out_cached": 0.1, "context_window": 196608, "default_max_tokens": 19660, "can_reason": true, @@ -201,10 +201,10 @@ { "id": "hf:MiniMaxAI/MiniMax-M2.5", "name": "MiniMax M2.5", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 3, - "cost_per_1m_in_cached": 0.6, - "cost_per_1m_out_cached": 0.6, + "cost_per_1m_in": 0.4, + "cost_per_1m_out": 2, + "cost_per_1m_in_cached": 0.4, + "cost_per_1m_out_cached": 0.4, "context_window": 191488, "default_max_tokens": 19148, "can_reason": true, @@ -220,10 +220,10 @@ { "id": "hf:nvidia/NVIDIA-Nemotron-3-Super-120B-A12B-NVFP4", "name": "NVIDIA Nemotron 3 Super 120B A12B NVFP4", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 3, - "cost_per_1m_in_cached": 0.6, - "cost_per_1m_out_cached": 0.6, + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 1, + "cost_per_1m_in_cached": 0.3, + "cost_per_1m_out_cached": 0.3, "context_window": 262144, "default_max_tokens": 32768, "can_reason": true, diff --git a/internal/providers/configs/venice.json b/internal/providers/configs/venice.json index 277e5691e8a7346f6be165a19a1417d03e2865d6..ebaf6eade0f561730366b499b33e4ae57369edac 100644 --- a/internal/providers/configs/venice.json +++ b/internal/providers/configs/venice.json @@ -522,6 +522,25 @@ "supports_attachments": false, "options": {} }, + { + "id": "minimax-m27", + "name": "MiniMax M2.7", + "cost_per_1m_in": 0.375, + "cost_per_1m_out": 1.5, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 198000, + "default_max_tokens": 32768, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false, + "options": {} + }, { "id": "mistral-small-3-2-24b-instruct", "name": "Mistral Small 3.2 24B Instruct", diff --git a/internal/providers/configs/vercel.json b/internal/providers/configs/vercel.json index fd1df703b53335c91313fc59696211d15820eed1..0ba10df206c667d9ce7686666c98d61a9721eb0c 100644 --- a/internal/providers/configs/vercel.json +++ b/internal/providers/configs/vercel.json @@ -1852,6 +1852,44 @@ "supports_attachments": false, "options": {} }, + { + "id": "minimax/minimax-m2.7-highspeed", + "name": "MiniMax M2.7 High Speed", + "cost_per_1m_in": 0.6, + "cost_per_1m_out": 2.4, + "cost_per_1m_in_cached": 0.06, + "cost_per_1m_out_cached": 0.375, + "context_window": 204800, + "default_max_tokens": 8000, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": true, + "options": {} + }, + { + "id": "minimax/minimax-m2.7", + "name": "Minimax M2.7", + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 1.2, + "cost_per_1m_in_cached": 0.06, + "cost_per_1m_out_cached": 0.375, + "context_window": 204800, + "default_max_tokens": 8000, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": true, + "options": {} + }, { "id": "mistral/ministral-3b", "name": "Ministral 3B",