diff --git a/internal/providers/configs/aihubmix.json b/internal/providers/configs/aihubmix.json index d485b262ac6a97293ec8b5fa60764774ab8641be..0e124f99c6ca428a95f1e8d887cc5a920f9dc433 100644 --- a/internal/providers/configs/aihubmix.json +++ b/internal/providers/configs/aihubmix.json @@ -2466,6 +2466,32 @@ "supports_attachments": false, "options": {} }, + { + "id": "mimo-v2-omni", + "name": "MiMo V2 Omni", + "cost_per_1m_in": 0.44, + "cost_per_1m_out": 2.2, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0.088, + "context_window": 256000, + "default_max_tokens": 25600, + "can_reason": false, + "supports_attachments": true, + "options": {} + }, + { + "id": "mimo-v2-pro", + "name": "MiMo V2 Pro", + "cost_per_1m_in": 1.1, + "cost_per_1m_out": 3.3, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0.22, + "context_window": 1000000, + "default_max_tokens": 100000, + "can_reason": false, + "supports_attachments": true, + "options": {} + }, { "id": "minimax-m2", "name": "MiniMax M2", diff --git a/internal/providers/configs/huggingface.json b/internal/providers/configs/huggingface.json index 38daf7e3ef1faf6c3999b89dd2c7de35891a68e1..8628c41a762793b6caa50910815ef81fcbc1cd1a 100644 --- a/internal/providers/configs/huggingface.json +++ b/internal/providers/configs/huggingface.json @@ -20,6 +20,19 @@ "supports_attachments": false, "options": {} }, + { + "id": "MiniMaxAI/MiniMax-M2.5:fireworks-ai", + "name": "MiniMaxAI/MiniMax-M2.5 (fireworks-ai)", + "cost_per_1m_in": 0, + "cost_per_1m_out": 0, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 204800, + "default_max_tokens": 8192, + "can_reason": false, + "supports_attachments": false, + "options": {} + }, { "id": "Qwen/Qwen3-32B:groq", "name": "Qwen/Qwen3-32B (groq)", @@ -98,6 +111,19 @@ "supports_attachments": false, "options": {} }, + { + "id": "deepseek-ai/DeepSeek-V3.2:fireworks-ai", + "name": "deepseek-ai/DeepSeek-V3.2 (fireworks-ai)", + "cost_per_1m_in": 0, + "cost_per_1m_out": 0, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 163840, + "default_max_tokens": 8192, + "can_reason": false, + "supports_attachments": false, + "options": {} + }, { "id": "meta-llama/Llama-3.3-70B-Instruct:groq", "name": "meta-llama/Llama-3.3-70B-Instruct (groq)", @@ -137,19 +163,6 @@ "supports_attachments": false, "options": {} }, - { - "id": "moonshotai/Kimi-K2-Instruct-0905:groq", - "name": "moonshotai/Kimi-K2-Instruct-0905 (groq)", - "cost_per_1m_in": 0, - "cost_per_1m_out": 0, - "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 262144, - "default_max_tokens": 8192, - "can_reason": false, - "supports_attachments": false, - "options": {} - }, { "id": "moonshotai/Kimi-K2-Thinking:fireworks-ai", "name": "moonshotai/Kimi-K2-Thinking (fireworks-ai)", diff --git a/internal/providers/configs/openrouter.json b/internal/providers/configs/openrouter.json index 2352f2835d5b7383f5b55480db94e0e402f00e0f..264f49a7cbb4f9e55d299c5a6ba2a76ab1dd9c62 100644 --- a/internal/providers/configs/openrouter.json +++ b/internal/providers/configs/openrouter.json @@ -560,12 +560,12 @@ { "id": "deepseek/deepseek-chat-v3.1", "name": "DeepSeek: DeepSeek V3.1", - "cost_per_1m_in": 0.21, - "cost_per_1m_out": 0.79, + "cost_per_1m_in": 0.2, + "cost_per_1m_out": 0.8, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.13, + "cost_per_1m_out_cached": 0.1, "context_window": 163840, - "default_max_tokens": 16384, + "default_max_tokens": 32768, "can_reason": true, "reasoning_levels": [ "low", @@ -579,10 +579,10 @@ { "id": "deepseek/deepseek-v3.1-terminus", "name": "DeepSeek: DeepSeek V3.1 Terminus", - "cost_per_1m_in": 0.23, - "cost_per_1m_out": 0.9, + "cost_per_1m_in": 0.21, + "cost_per_1m_out": 0.78, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.115, + "cost_per_1m_out_cached": 0.105, "context_window": 163840, "default_max_tokens": 32768, "can_reason": true, @@ -603,7 +603,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.13, "context_window": 163840, - "default_max_tokens": 16384, + "default_max_tokens": 81920, "can_reason": true, "reasoning_levels": [ "low", @@ -998,7 +998,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 13107, + "default_max_tokens": 8192, "can_reason": false, "supports_attachments": false, "options": {} @@ -1006,12 +1006,12 @@ { "id": "meta-llama/llama-3.3-70b-instruct", "name": "Meta: Llama 3.3 70B Instruct", - "cost_per_1m_in": 0.25, - "cost_per_1m_out": 0.75, + "cost_per_1m_in": 0.13, + "cost_per_1m_out": 0.4, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 13107, + "default_max_tokens": 64000, "can_reason": false, "supports_attachments": false, "options": {} @@ -1078,7 +1078,7 @@ "id": "minimax/minimax-m2.1", "name": "MiniMax: MiniMax M2.1", "cost_per_1m_in": 0.3, - "cost_per_1m_out": 1.2, + "cost_per_1m_out": 2.4, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.03, "context_window": 204800, @@ -1096,12 +1096,12 @@ { "id": "minimax/minimax-m2.5", "name": "MiniMax: MiniMax M2.5", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 2.4, + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 1.2, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.06, + "cost_per_1m_out_cached": 0.03, "context_window": 204800, - "default_max_tokens": 65536, + "default_max_tokens": 65550, "can_reason": true, "reasoning_levels": [ "low", @@ -1134,8 +1134,8 @@ { "id": "minimax/minimax-m2.7", "name": "MiniMax: MiniMax M2.7", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 2.4, + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 1.2, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.06, "context_window": 204800, @@ -1484,12 +1484,12 @@ { "id": "moonshotai/kimi-k2-0905", "name": "MoonshotAI: Kimi K2 0905", - "cost_per_1m_in": 1, - "cost_per_1m_out": 3, + "cost_per_1m_in": 0.6, + "cost_per_1m_out": 2.5, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.5, + "cost_per_1m_out_cached": 0.6, "context_window": 262144, - "default_max_tokens": 8192, + "default_max_tokens": 16384, "can_reason": false, "supports_attachments": false, "options": {} @@ -1500,7 +1500,7 @@ "cost_per_1m_in": 0.6, "cost_per_1m_out": 2.5, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.15, "context_window": 262144, "default_max_tokens": 131072, "can_reason": true, @@ -1517,11 +1517,11 @@ "id": "moonshotai/kimi-k2.5", "name": "MoonshotAI: Kimi K2.5", "cost_per_1m_in": 0.6, - "cost_per_1m_out": 2.8, + "cost_per_1m_out": 3, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.1, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 26214, "can_reason": true, "reasoning_levels": [ "low", @@ -1602,6 +1602,25 @@ "supports_attachments": false, "options": {} }, + { + "id": "nvidia/nemotron-3-super-120b-a12b", + "name": "NVIDIA: Nemotron 3 Super", + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.5, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0.04, + "context_window": 262144, + "default_max_tokens": 26214, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false, + "options": {} + }, { "id": "nvidia/nemotron-3-super-120b-a12b:free", "name": "NVIDIA: Nemotron 3 Super (free)", @@ -1775,7 +1794,7 @@ "cost_per_1m_in": 2.5, "cost_per_1m_out": 10, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 1.25, + "cost_per_1m_out_cached": 0, "context_window": 128000, "default_max_tokens": 8192, "can_reason": false, @@ -2012,7 +2031,7 @@ "cost_per_1m_in": 1.25, "cost_per_1m_out": 10, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.125, + "cost_per_1m_out_cached": 0.13, "context_window": 400000, "default_max_tokens": 64000, "can_reason": true, @@ -2276,12 +2295,12 @@ { "id": "openai/gpt-oss-120b", "name": "OpenAI: gpt-oss-120b", - "cost_per_1m_in": 0.05, - "cost_per_1m_out": 0.45, + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.49, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.025, + "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 32768, + "default_max_tokens": 13107, "can_reason": true, "reasoning_levels": [ "low", @@ -2687,6 +2706,25 @@ "supports_attachments": false, "options": {} }, + { + "id": "qwen/qwen3-14b", + "name": "Qwen: Qwen3 14B", + "cost_per_1m_in": 0.2275, + "cost_per_1m_out": 0.91, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 131072, + "default_max_tokens": 4096, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false, + "options": {} + }, { "id": "qwen/qwen3-235b-a22b", "name": "Qwen: Qwen3 235B A22B", @@ -2709,8 +2747,8 @@ { "id": "qwen/qwen3-235b-a22b-2507", "name": "Qwen: Qwen3 235B A22B Instruct 2507", - "cost_per_1m_in": 0.1, - "cost_per_1m_out": 0.1, + "cost_per_1m_in": 0.2, + "cost_per_1m_out": 0.8, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, @@ -2862,12 +2900,12 @@ { "id": "qwen/qwen3-coder", "name": "Qwen: Qwen3 Coder 480B A35B", - "cost_per_1m_in": 0.22, + "cost_per_1m_in": 0.4, "cost_per_1m_out": 1.8, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 32768, + "default_max_tokens": 131072, "can_reason": false, "supports_attachments": false, "options": {} @@ -2959,12 +2997,12 @@ { "id": "qwen/qwen3-next-80b-a3b-instruct", "name": "Qwen: Qwen3 Next 80B A3B Instruct", - "cost_per_1m_in": 0.09, + "cost_per_1m_in": 0.1, "cost_per_1m_out": 1.1, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 26214, + "default_max_tokens": 131072, "can_reason": false, "supports_attachments": false, "options": {} @@ -3049,12 +3087,12 @@ { "id": "qwen/qwen3-vl-30b-a3b-thinking", "name": "Qwen: Qwen3 VL 30B A3B Thinking", - "cost_per_1m_in": 0.2, + "cost_per_1m_in": 0.29, "cost_per_1m_out": 1, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 131072, - "default_max_tokens": 16384, + "context_window": 262144, + "default_max_tokens": 131072, "can_reason": true, "reasoning_levels": [ "low", @@ -3151,10 +3189,10 @@ { "id": "qwen/qwen3.5-122b-a10b", "name": "Qwen: Qwen3.5-122B-A10B", - "cost_per_1m_in": 0.4, - "cost_per_1m_out": 3.2, + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 2.4, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.3, "context_window": 262144, "default_max_tokens": 32768, "can_reason": true, @@ -3453,9 +3491,9 @@ "cost_per_1m_in": 0.6, "cost_per_1m_out": 2.2, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.11, + "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 49152, + "default_max_tokens": 13107, "can_reason": true, "reasoning_levels": [ "low", @@ -3548,9 +3586,9 @@ "cost_per_1m_in": 0.3, "cost_per_1m_out": 0.9, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.05, "context_window": 131072, - "default_max_tokens": 65536, + "default_max_tokens": 12000, "can_reason": true, "reasoning_levels": [ "low", @@ -3564,12 +3602,12 @@ { "id": "z-ai/glm-4.7", "name": "Z.ai: GLM 4.7", - "cost_per_1m_in": 0.44, - "cost_per_1m_out": 2, + "cost_per_1m_in": 0.54, + "cost_per_1m_out": 1.98, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.22, - "context_window": 202752, - "default_max_tokens": 20275, + "cost_per_1m_out_cached": 0.099, + "context_window": 204800, + "default_max_tokens": 65536, "can_reason": true, "reasoning_levels": [ "low", @@ -3583,10 +3621,10 @@ { "id": "z-ai/glm-4.7-flash", "name": "Z.ai: GLM 4.7 Flash", - "cost_per_1m_in": 0.06, - "cost_per_1m_out": 0.4, + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.43, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.01, + "cost_per_1m_out_cached": 0, "context_window": 202752, "default_max_tokens": 20275, "can_reason": true, @@ -3602,11 +3640,11 @@ { "id": "z-ai/glm-5", "name": "Z.ai: GLM 5", - "cost_per_1m_in": 0.95, - "cost_per_1m_out": 2.55, + "cost_per_1m_in": 1, + "cost_per_1m_out": 3.2, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.2, - "context_window": 204800, + "context_window": 202800, "default_max_tokens": 65536, "can_reason": true, "reasoning_levels": [