diff --git a/internal/providers/configs/huggingface.json b/internal/providers/configs/huggingface.json index 159ee9f385d27c2954305f68d71c143eb37714ac..9d4189a85dc7fb887bc7e572944703b65805a999 100644 --- a/internal/providers/configs/huggingface.json +++ b/internal/providers/configs/huggingface.json @@ -150,6 +150,19 @@ "supports_attachments": false, "options": {} }, + { + "id": "moonshotai/Kimi-K2-Instruct-0905:fireworks-ai", + "name": "moonshotai/Kimi-K2-Instruct-0905 (fireworks-ai)", + "cost_per_1m_in": 0, + "cost_per_1m_out": 0, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 262144, + "default_max_tokens": 8192, + "can_reason": false, + "supports_attachments": false, + "options": {} + }, { "id": "moonshotai/Kimi-K2-Instruct-0905:groq", "name": "moonshotai/Kimi-K2-Instruct-0905 (groq)", diff --git a/internal/providers/configs/openrouter.json b/internal/providers/configs/openrouter.json index f0717dc64eff714dc98335fa4517afaff2d0746f..a175c1e8e73183dec4e1eacf51e7e8e2a22f077d 100644 --- a/internal/providers/configs/openrouter.json +++ b/internal/providers/configs/openrouter.json @@ -135,8 +135,8 @@ "name": "Anthropic: Claude 3.5 Sonnet", "cost_per_1m_in": 6, "cost_per_1m_out": 30, - "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_in_cached": 7.5, + "cost_per_1m_out_cached": 0.6, "context_window": 200000, "default_max_tokens": 4096, "can_reason": false, @@ -560,10 +560,10 @@ { "id": "deepseek/deepseek-chat-v3.1", "name": "DeepSeek: DeepSeek V3.1", - "cost_per_1m_in": 0.21, - "cost_per_1m_out": 0.79, + "cost_per_1m_in": 0.6, + "cost_per_1m_out": 1.7, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.13, + "cost_per_1m_out_cached": 0, "context_window": 163840, "default_max_tokens": 16384, "can_reason": true, @@ -598,12 +598,12 @@ { "id": "deepseek/deepseek-v3.2", "name": "DeepSeek: DeepSeek V3.2", - "cost_per_1m_in": 0.56, - "cost_per_1m_out": 1.68, + "cost_per_1m_in": 0.26, + "cost_per_1m_out": 0.38, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.13, "context_window": 163840, - "default_max_tokens": 32768, + "default_max_tokens": 81920, "can_reason": true, "reasoning_levels": [ "low", @@ -832,7 +832,7 @@ "cost_per_1m_in_cached": 0.08333, "cost_per_1m_out_cached": 0.05, "context_window": 1048576, - "default_max_tokens": 32767, + "default_max_tokens": 32768, "can_reason": true, "reasoning_levels": [ "low", @@ -1006,12 +1006,12 @@ { "id": "meta-llama/llama-3.3-70b-instruct", "name": "Meta: Llama 3.3 70B Instruct", - "cost_per_1m_in": 0.13, + "cost_per_1m_in": 0.135, "cost_per_1m_out": 0.4, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 64000, + "default_max_tokens": 60000, "can_reason": false, "supports_attachments": false, "options": {} @@ -1096,10 +1096,10 @@ { "id": "minimax/minimax-m2.5", "name": "MiniMax: MiniMax M2.5", - "cost_per_1m_in": 0.3, - "cost_per_1m_out": 1.2, + "cost_per_1m_in": 0.6, + "cost_per_1m_out": 2.4, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.03, + "cost_per_1m_out_cached": 0.06, "context_window": 204800, "default_max_tokens": 65536, "can_reason": true, @@ -1134,8 +1134,8 @@ { "id": "minimax/minimax-m2.7", "name": "MiniMax: MiniMax M2.7", - "cost_per_1m_in": 0.3, - "cost_per_1m_out": 1.2, + "cost_per_1m_in": 0.6, + "cost_per_1m_out": 2.4, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.06, "context_window": 204800, @@ -1247,9 +1247,9 @@ "cost_per_1m_in": 0.2, "cost_per_1m_out": 0.2, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.02, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 26214, "can_reason": false, "supports_attachments": true, "options": {} @@ -1273,9 +1273,9 @@ "cost_per_1m_in": 0.15, "cost_per_1m_out": 0.15, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.015, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 26214, "can_reason": false, "supports_attachments": true, "options": {} @@ -1361,12 +1361,12 @@ { "id": "mistralai/mistral-small-3.2-24b-instruct", "name": "Mistral: Mistral Small 3.2 24B", - "cost_per_1m_in": 0.1, - "cost_per_1m_out": 0.3, + "cost_per_1m_in": 0.09375, + "cost_per_1m_out": 0.25, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.01, - "context_window": 131072, - "default_max_tokens": 13107, + "cost_per_1m_out_cached": 0, + "context_window": 256000, + "default_max_tokens": 8192, "can_reason": false, "supports_attachments": true, "options": {} @@ -1487,9 +1487,9 @@ "cost_per_1m_in": 0.6, "cost_per_1m_out": 2.5, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.6, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 16384, "can_reason": false, "supports_attachments": false, "options": {} @@ -1502,7 +1502,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 26214, + "default_max_tokens": 131072, "can_reason": true, "reasoning_levels": [ "low", @@ -1608,7 +1608,7 @@ "cost_per_1m_in": 0.1, "cost_per_1m_out": 0.5, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.04, + "cost_per_1m_out_cached": 0.1, "context_window": 262144, "default_max_tokens": 26214, "can_reason": true, @@ -1770,7 +1770,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.1, "context_window": 1047576, - "default_max_tokens": 16384, + "default_max_tokens": 104757, "can_reason": false, "supports_attachments": true, "options": {} @@ -2296,11 +2296,11 @@ "id": "openai/gpt-oss-120b", "name": "OpenAI: gpt-oss-120b", "cost_per_1m_in": 0.05, - "cost_per_1m_out": 0.45, + "cost_per_1m_out": 0.25, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.025, + "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 32768, + "default_max_tokens": 16384, "can_reason": true, "reasoning_levels": [ "low", @@ -2706,6 +2706,25 @@ "supports_attachments": false, "options": {} }, + { + "id": "qwen/qwen3-14b", + "name": "Qwen: Qwen3 14B", + "cost_per_1m_in": 0.12, + "cost_per_1m_out": 0.24, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 40960, + "default_max_tokens": 20480, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false, + "options": {} + }, { "id": "qwen/qwen3-235b-a22b", "name": "Qwen: Qwen3 235B A22B", @@ -2792,12 +2811,12 @@ { "id": "qwen/qwen3-30b-a3b-thinking-2507", "name": "Qwen: Qwen3 30B A3B Thinking 2507", - "cost_per_1m_in": 0.1, - "cost_per_1m_out": 0.3, + "cost_per_1m_in": 0.08, + "cost_per_1m_out": 0.4, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 262144, - "default_max_tokens": 26214, + "cost_per_1m_out_cached": 0.08, + "context_window": 131072, + "default_max_tokens": 65536, "can_reason": true, "reasoning_levels": [ "low", @@ -2868,12 +2887,12 @@ { "id": "qwen/qwen3-coder-30b-a3b-instruct", "name": "Qwen: Qwen3 Coder 30B A3B Instruct", - "cost_per_1m_in": 0.07, - "cost_per_1m_out": 0.27, + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.3, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 160000, - "default_max_tokens": 16384, + "context_window": 262144, + "default_max_tokens": 26214, "can_reason": false, "supports_attachments": false, "options": {} @@ -3023,12 +3042,12 @@ { "id": "qwen/qwen3-vl-235b-a22b-instruct", "name": "Qwen: Qwen3 VL 235B A22B Instruct", - "cost_per_1m_in": 0.25, - "cost_per_1m_out": 1.5, + "cost_per_1m_in": 0.21, + "cost_per_1m_out": 1.9, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 256000, - "default_max_tokens": 25600, + "context_window": 131072, + "default_max_tokens": 16384, "can_reason": false, "supports_attachments": true, "options": {} @@ -3068,12 +3087,12 @@ { "id": "qwen/qwen3-vl-30b-a3b-thinking", "name": "Qwen: Qwen3 VL 30B A3B Thinking", - "cost_per_1m_in": 0.2, + "cost_per_1m_in": 0.29, "cost_per_1m_out": 1, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 131072, - "default_max_tokens": 16384, + "context_window": 262144, + "default_max_tokens": 131072, "can_reason": true, "reasoning_levels": [ "low", @@ -3189,10 +3208,10 @@ { "id": "qwen/qwen3.5-27b", "name": "Qwen: Qwen3.5-27B", - "cost_per_1m_in": 0.3, - "cost_per_1m_out": 2.4, + "cost_per_1m_in": 0.27, + "cost_per_1m_out": 2.16, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.27, "context_window": 262144, "default_max_tokens": 32768, "can_reason": true, @@ -3227,12 +3246,12 @@ { "id": "qwen/qwen3.5-9b", "name": "Qwen: Qwen3.5-9B", - "cost_per_1m_in": 0.05, + "cost_per_1m_in": 0.1, "cost_per_1m_out": 0.15, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 256000, - "default_max_tokens": 25600, + "context_window": 262144, + "default_max_tokens": 26214, "can_reason": true, "reasoning_levels": [ "low", @@ -3488,12 +3507,12 @@ { "id": "z-ai/glm-4.5-air", "name": "Z.ai: GLM 4.5 Air", - "cost_per_1m_in": 0.13, - "cost_per_1m_out": 0.85, + "cost_per_1m_in": 0.2, + "cost_per_1m_out": 1.2, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.025, + "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 49152, + "default_max_tokens": 13107, "can_reason": true, "reasoning_levels": [ "low", @@ -3545,12 +3564,12 @@ { "id": "z-ai/glm-4.6", "name": "Z.ai: GLM 4.6", - "cost_per_1m_in": 0.55, - "cost_per_1m_out": 2.2, + "cost_per_1m_in": 0.39, + "cost_per_1m_out": 1.9, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.11, + "cost_per_1m_out_cached": 0, "context_window": 204800, - "default_max_tokens": 65536, + "default_max_tokens": 102400, "can_reason": true, "reasoning_levels": [ "low", @@ -3602,10 +3621,10 @@ { "id": "z-ai/glm-4.7-flash", "name": "Z.ai: GLM 4.7 Flash", - "cost_per_1m_in": 0.1, - "cost_per_1m_out": 0.43, + "cost_per_1m_in": 0.06, + "cost_per_1m_out": 0.4, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.01, "context_window": 202752, "default_max_tokens": 20275, "can_reason": true, @@ -3622,11 +3641,11 @@ "id": "z-ai/glm-5", "name": "Z.ai: GLM 5", "cost_per_1m_in": 0.95, - "cost_per_1m_out": 2.55, + "cost_per_1m_out": 3.15, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.2, - "context_window": 204800, - "default_max_tokens": 65536, + "cost_per_1m_out_cached": 0, + "context_window": 202800, + "default_max_tokens": 101400, "can_reason": true, "reasoning_levels": [ "low", @@ -3640,10 +3659,10 @@ { "id": "z-ai/glm-5-turbo", "name": "Z.ai: GLM 5 Turbo", - "cost_per_1m_in": 0.96, - "cost_per_1m_out": 3.2, + "cost_per_1m_in": 1.2, + "cost_per_1m_out": 4, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.192, + "cost_per_1m_out_cached": 0.24, "context_window": 202752, "default_max_tokens": 65536, "can_reason": true, diff --git a/internal/providers/configs/vercel.json b/internal/providers/configs/vercel.json index a0fe6d912e538fe27e5f7cb549ef5d18bb123b33..7f82a70e21acaee5e444317a068bdfa569fe1dc2 100644 --- a/internal/providers/configs/vercel.json +++ b/internal/providers/configs/vercel.json @@ -1449,6 +1449,57 @@ "supports_attachments": false, "options": {} }, + { + "id": "xai/grok-4.20-multi-agent", + "name": "Grok 4.20 Multi-Agent", + "cost_per_1m_in": 2, + "cost_per_1m_out": 6, + "cost_per_1m_in_cached": 0.2, + "cost_per_1m_out_cached": 0, + "context_window": 2000000, + "default_max_tokens": 8000, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false, + "options": {} + }, + { + "id": "xai/grok-4.20-non-reasoning", + "name": "Grok 4.20 Non-Reasoning", + "cost_per_1m_in": 2, + "cost_per_1m_out": 6, + "cost_per_1m_in_cached": 0.2, + "cost_per_1m_out_cached": 0, + "context_window": 2000000, + "default_max_tokens": 8000, + "can_reason": false, + "supports_attachments": true, + "options": {} + }, + { + "id": "xai/grok-4.20-reasoning", + "name": "Grok 4.20 Reasoning", + "cost_per_1m_in": 2, + "cost_per_1m_out": 6, + "cost_per_1m_in_cached": 0.2, + "cost_per_1m_out_cached": 0, + "context_window": 2000000, + "default_max_tokens": 8000, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": true, + "options": {} + }, { "id": "xai/grok-code-fast-1", "name": "Grok Code Fast 1", @@ -2309,8 +2360,8 @@ { "id": "perplexity/sonar", "name": "Sonar", - "cost_per_1m_in": 1, - "cost_per_1m_out": 1, + "cost_per_1m_in": 0, + "cost_per_1m_out": 0, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 127000, @@ -2322,8 +2373,8 @@ { "id": "perplexity/sonar-pro", "name": "Sonar Pro", - "cost_per_1m_in": 3, - "cost_per_1m_out": 15, + "cost_per_1m_in": 0, + "cost_per_1m_out": 0, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 200000,