diff --git a/internal/providers/configs/openrouter.json b/internal/providers/configs/openrouter.json index 3546e34eb0acf70a4b754956a62e17359800b493..89d0978a80dd3a22df3a843d1e60d439a382c5db 100644 --- a/internal/providers/configs/openrouter.json +++ b/internal/providers/configs/openrouter.json @@ -46,6 +46,19 @@ "supports_attachments": false, "options": {} }, + { + "id": "allenai/olmo-3.1-32b-instruct", + "name": "AllenAI: Olmo 3.1 32B Instruct", + "cost_per_1m_in": 0.19999999999999998, + "cost_per_1m_out": 0.6, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 65536, + "default_max_tokens": 6553, + "can_reason": false, + "supports_attachments": false, + "options": {} + }, { "id": "amazon/nova-2-lite-v1", "name": "Amazon: Nova 2 Lite", @@ -177,7 +190,7 @@ "cost_per_1m_in_cached": 3.75, "cost_per_1m_out_cached": 0.3, "context_window": 200000, - "default_max_tokens": 64000, + "default_max_tokens": 32000, "can_reason": true, "reasoning_levels": [ "low", @@ -560,12 +573,12 @@ { "id": "deepseek/deepseek-chat-v3.1", "name": "DeepSeek: DeepSeek V3.1", - "cost_per_1m_in": 0.19999999999999998, - "cost_per_1m_out": 0.7999999999999999, + "cost_per_1m_in": 0.21, + "cost_per_1m_out": 0.7899999999999999, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.16799999999999998, "context_window": 163840, - "default_max_tokens": 32768, + "default_max_tokens": 16384, "can_reason": true, "reasoning_levels": [ "low", @@ -599,10 +612,10 @@ "id": "deepseek/deepseek-v3.1-terminus:exacto", "name": "DeepSeek: DeepSeek V3.1 Terminus (exacto)", "cost_per_1m_in": 0.21, - "cost_per_1m_out": 0.7899999999999999, + "cost_per_1m_out": 0.7999999999999999, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.16799999999999998, - "context_window": 163840, + "cost_per_1m_out_cached": 0, + "context_window": 131072, "default_max_tokens": 16384, "can_reason": true, "reasoning_levels": [ @@ -617,12 +630,12 @@ { "id": "deepseek/deepseek-v3.2", "name": "DeepSeek: DeepSeek V3.2", - "cost_per_1m_in": 0.26, - "cost_per_1m_out": 0.39, + "cost_per_1m_in": 0.56, + "cost_per_1m_out": 1.68, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.13, + "cost_per_1m_out_cached": 0, "context_window": 163840, - "default_max_tokens": 16384, + "default_max_tokens": 32768, "can_reason": true, "reasoning_levels": [ "low", @@ -813,7 +826,7 @@ "cost_per_1m_in_cached": 0.3833, "cost_per_1m_out_cached": 0.075, "context_window": 1048576, - "default_max_tokens": 32768, + "default_max_tokens": 32767, "can_reason": true, "reasoning_levels": [ "low", @@ -1282,6 +1295,19 @@ "supports_attachments": false, "options": {} }, + { + "id": "mistralai/mistral-7b-instruct", + "name": "Mistral: Mistral 7B Instruct", + "cost_per_1m_in": 0.028, + "cost_per_1m_out": 0.054, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 32768, + "default_max_tokens": 8192, + "can_reason": false, + "supports_attachments": false, + "options": {} + }, { "id": "mistralai/mistral-7b-instruct:free", "name": "Mistral: Mistral 7B Instruct (free)", @@ -1493,12 +1519,12 @@ { "id": "moonshotai/kimi-k2-0905:exacto", "name": "MoonshotAI: Kimi K2 0905 (exacto)", - "cost_per_1m_in": 1, - "cost_per_1m_out": 3, + "cost_per_1m_in": 0.6, + "cost_per_1m_out": 2.5, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 8192, + "default_max_tokens": 26214, "can_reason": false, "supports_attachments": false, "options": {} @@ -1506,10 +1532,10 @@ { "id": "moonshotai/kimi-k2-thinking", "name": "MoonshotAI: Kimi K2 Thinking", - "cost_per_1m_in": 0.6, + "cost_per_1m_in": 0.55, "cost_per_1m_out": 2.5, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.15, + "cost_per_1m_out_cached": 0, "context_window": 262144, "default_max_tokens": 131072, "can_reason": true, @@ -1562,7 +1588,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 26214, + "default_max_tokens": 131072, "can_reason": true, "reasoning_levels": [ "low", @@ -1681,25 +1707,6 @@ "supports_attachments": false, "options": {} }, - { - "id": "nousresearch/hermes-4-70b", - "name": "Nous: Hermes 4 70B", - "cost_per_1m_in": 0.11, - "cost_per_1m_out": 0.38, - "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 131072, - "default_max_tokens": 65536, - "can_reason": true, - "reasoning_levels": [ - "low", - "medium", - "high" - ], - "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} - }, { "id": "openai/codex-mini", "name": "OpenAI: Codex Mini", @@ -1779,7 +1786,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.09999999999999999, "context_window": 1047576, - "default_max_tokens": 104757, + "default_max_tokens": 16384, "can_reason": false, "supports_attachments": true, "options": {} @@ -1790,9 +1797,9 @@ "cost_per_1m_in": 0.09999999999999999, "cost_per_1m_out": 0.39999999999999997, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.03, + "cost_per_1m_out_cached": 0.024999999999999998, "context_window": 1047576, - "default_max_tokens": 104757, + "default_max_tokens": 16384, "can_reason": false, "supports_attachments": true, "options": {} @@ -2002,9 +2009,9 @@ "cost_per_1m_in": 0.049999999999999996, "cost_per_1m_out": 0.39999999999999997, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.005, + "cost_per_1m_out_cached": 0.01, "context_window": 400000, - "default_max_tokens": 64000, + "default_max_tokens": 40000, "can_reason": true, "reasoning_levels": [ "low", @@ -2177,8 +2184,8 @@ { "id": "openai/gpt-oss-120b", "name": "OpenAI: gpt-oss-120b", - "cost_per_1m_in": 0.02, - "cost_per_1m_out": 0.09999999999999999, + "cost_per_1m_in": 0.039, + "cost_per_1m_out": 0.19, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, @@ -2234,12 +2241,12 @@ { "id": "openai/gpt-oss-20b", "name": "OpenAI: gpt-oss-20b", - "cost_per_1m_in": 0.02, - "cost_per_1m_out": 0.09999999999999999, + "cost_per_1m_in": 0.049999999999999996, + "cost_per_1m_out": 0.19999999999999998, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 65536, + "default_max_tokens": 13107, "can_reason": true, "reasoning_levels": [ "low", @@ -2461,21 +2468,15 @@ "options": {} }, { - "id": "qwen/qwq-32b", - "name": "Qwen: QwQ 32B", - "cost_per_1m_in": 0.15, - "cost_per_1m_out": 0.58, + "id": "qwen/qwen-2.5-72b-instruct", + "name": "Qwen2.5 72B Instruct", + "cost_per_1m_in": 0.12, + "cost_per_1m_out": 0.39, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 131072, - "default_max_tokens": 65536, - "can_reason": true, - "reasoning_levels": [ - "low", - "medium", - "high" - ], - "default_reasoning_effort": "medium", + "context_window": 32768, + "default_max_tokens": 8192, + "can_reason": false, "supports_attachments": false, "options": {} }, @@ -2563,6 +2564,19 @@ "supports_attachments": false, "options": {} }, + { + "id": "qwen/qwen-2.5-7b-instruct", + "name": "Qwen: Qwen2.5 7B Instruct", + "cost_per_1m_in": 0.04, + "cost_per_1m_out": 0.09999999999999999, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 128000, + "default_max_tokens": 4096, + "can_reason": false, + "supports_attachments": false, + "options": {} + }, { "id": "qwen/qwen3-14b", "name": "Qwen: Qwen3 14B", @@ -2665,6 +2679,25 @@ "supports_attachments": false, "options": {} }, + { + "id": "qwen/qwen3-30b-a3b-thinking-2507", + "name": "Qwen: Qwen3 30B A3B Thinking 2507", + "cost_per_1m_in": 0.09, + "cost_per_1m_out": 0.3, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 262144, + "default_max_tokens": 65536, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false, + "options": {} + }, { "id": "qwen/qwen3-32b", "name": "Qwen: Qwen3 32B", @@ -2703,6 +2736,25 @@ "supports_attachments": false, "options": {} }, + { + "id": "qwen/qwen3-8b", + "name": "Qwen: Qwen3 8B", + "cost_per_1m_in": 0.049999999999999996, + "cost_per_1m_out": 0.25, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 32000, + "default_max_tokens": 4096, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false, + "options": {} + }, { "id": "qwen/qwen3-coder-30b-a3b-instruct", "name": "Qwen: Qwen3 Coder 30B A3B Instruct", @@ -2797,12 +2849,12 @@ { "id": "qwen/qwen3-next-80b-a3b-instruct", "name": "Qwen: Qwen3 Next 80B A3B Instruct", - "cost_per_1m_in": 0.06, - "cost_per_1m_out": 0.6, + "cost_per_1m_in": 0.14, + "cost_per_1m_out": 1.4, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 26214, + "default_max_tokens": 131072, "can_reason": false, "supports_attachments": false, "options": {} @@ -2810,8 +2862,8 @@ { "id": "qwen/qwen3-next-80b-a3b-thinking", "name": "Qwen: Qwen3 Next 80B A3B Thinking", - "cost_per_1m_in": 0.3, - "cost_per_1m_out": 0.3, + "cost_per_1m_in": 0.15, + "cost_per_1m_out": 1.2, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, @@ -2829,12 +2881,12 @@ { "id": "qwen/qwen3-vl-235b-a22b-instruct", "name": "Qwen: Qwen3 VL 235B A22B Instruct", - "cost_per_1m_in": 0.39999999999999997, - "cost_per_1m_out": 1.5999999999999999, + "cost_per_1m_in": 0.22, + "cost_per_1m_out": 0.88, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 131072, - "default_max_tokens": 16384, + "context_window": 262144, + "default_max_tokens": 26214, "can_reason": false, "supports_attachments": true, "options": {} @@ -3077,7 +3129,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.11, "context_window": 131072, - "default_max_tokens": 49152, + "default_max_tokens": 48000, "can_reason": true, "reasoning_levels": [ "low", @@ -3148,12 +3200,12 @@ { "id": "z-ai/glm-4.6", "name": "Z.AI: GLM 4.6", - "cost_per_1m_in": 0.55, - "cost_per_1m_out": 2.2, + "cost_per_1m_in": 0.39, + "cost_per_1m_out": 1.9, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.11, + "cost_per_1m_out_cached": 0, "context_window": 204800, - "default_max_tokens": 65536, + "default_max_tokens": 102400, "can_reason": true, "reasoning_levels": [ "low", @@ -3205,12 +3257,12 @@ { "id": "z-ai/glm-4.7", "name": "Z.AI: GLM 4.7", - "cost_per_1m_in": 0.54, - "cost_per_1m_out": 1.9800000000000002, + "cost_per_1m_in": 0.42, + "cost_per_1m_out": 2.2, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.099, + "cost_per_1m_out_cached": 0, "context_window": 204800, - "default_max_tokens": 65536, + "default_max_tokens": 102400, "can_reason": true, "reasoning_levels": [ "low", @@ -3224,10 +3276,10 @@ { "id": "x-ai/grok-3", "name": "xAI: Grok 3", - "cost_per_1m_in": 5, - "cost_per_1m_out": 25, + "cost_per_1m_in": 3, + "cost_per_1m_out": 15, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 1.25, + "cost_per_1m_out_cached": 0.75, "context_window": 131072, "default_max_tokens": 13107, "can_reason": false, @@ -3237,10 +3289,10 @@ { "id": "x-ai/grok-3-beta", "name": "xAI: Grok 3 Beta", - "cost_per_1m_in": 5, - "cost_per_1m_out": 25, + "cost_per_1m_in": 3, + "cost_per_1m_out": 15, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 1.25, + "cost_per_1m_out_cached": 0.75, "context_window": 131072, "default_max_tokens": 13107, "can_reason": false,